Picture for Jianing Yang

Jianing Yang

Error-Driven Scene Editing for 3D Grounding in Large Language Models

Add code
Nov 18, 2025
Viaarxiv icon

Emotional Text-To-Speech Based on Mutual-Information-Guided Emotion-Timbre Disentanglement

Add code
Oct 02, 2025
Figure 1 for Emotional Text-To-Speech Based on Mutual-Information-Guided Emotion-Timbre Disentanglement
Figure 2 for Emotional Text-To-Speech Based on Mutual-Information-Guided Emotion-Timbre Disentanglement
Figure 3 for Emotional Text-To-Speech Based on Mutual-Information-Guided Emotion-Timbre Disentanglement
Figure 4 for Emotional Text-To-Speech Based on Mutual-Information-Guided Emotion-Timbre Disentanglement
Viaarxiv icon

4D-LRM: Large Space-Time Reconstruction Model From and To Any View at Any Time

Add code
Jun 23, 2025
Viaarxiv icon

LIFT-GS: Cross-Scene Render-Supervised Distillation for 3D Language Grounding

Add code
Feb 27, 2025
Viaarxiv icon

Fast3R: Towards 3D Reconstruction of 1000+ Images in One Forward Pass

Add code
Jan 23, 2025
Viaarxiv icon

Teaching Embodied Reinforcement Learning Agents: Informativeness and Diversity of Language Use

Add code
Oct 31, 2024
Viaarxiv icon

Multi-Object Hallucination in Vision-Language Models

Add code
Jul 08, 2024
Figure 1 for Multi-Object Hallucination in Vision-Language Models
Figure 2 for Multi-Object Hallucination in Vision-Language Models
Figure 3 for Multi-Object Hallucination in Vision-Language Models
Figure 4 for Multi-Object Hallucination in Vision-Language Models
Viaarxiv icon

Improving Autoregressive Training with Dynamic Oracles

Add code
Jun 13, 2024
Figure 1 for Improving Autoregressive Training with Dynamic Oracles
Figure 2 for Improving Autoregressive Training with Dynamic Oracles
Figure 3 for Improving Autoregressive Training with Dynamic Oracles
Figure 4 for Improving Autoregressive Training with Dynamic Oracles
Viaarxiv icon

3D-GRAND: A Million-Scale Dataset for 3D-LLMs with Better Grounding and Less Hallucination

Add code
Jun 12, 2024
Figure 1 for 3D-GRAND: A Million-Scale Dataset for 3D-LLMs with Better Grounding and Less Hallucination
Figure 2 for 3D-GRAND: A Million-Scale Dataset for 3D-LLMs with Better Grounding and Less Hallucination
Figure 3 for 3D-GRAND: A Million-Scale Dataset for 3D-LLMs with Better Grounding and Less Hallucination
Figure 4 for 3D-GRAND: A Million-Scale Dataset for 3D-LLMs with Better Grounding and Less Hallucination
Viaarxiv icon

3D-GRAND: Towards Better Grounding and Less Hallucination for 3D-LLMs

Add code
Jun 07, 2024
Figure 1 for 3D-GRAND: Towards Better Grounding and Less Hallucination for 3D-LLMs
Figure 2 for 3D-GRAND: Towards Better Grounding and Less Hallucination for 3D-LLMs
Figure 3 for 3D-GRAND: Towards Better Grounding and Less Hallucination for 3D-LLMs
Figure 4 for 3D-GRAND: Towards Better Grounding and Less Hallucination for 3D-LLMs
Viaarxiv icon