Picture for Yu-Xiong Wang

Yu-Xiong Wang

Can We Generate Visual Programs Without Prompting LLMs?

Add code
Dec 11, 2024
Viaarxiv icon

PaintScene4D: Consistent 4D Scene Generation from Text Prompts

Add code
Dec 05, 2024
Viaarxiv icon

RandAR: Decoder-only Autoregressive Visual Generation in Random Orders

Add code
Dec 02, 2024
Viaarxiv icon

Diff-2-in-1: Bridging Generation and Dense Perception with Diffusion Models

Add code
Nov 07, 2024
Viaarxiv icon

ProEdit: Simple Progression is All You Need for High-Quality 3D Scene Editing

Add code
Nov 07, 2024
Viaarxiv icon

Reinforcement Learning Gradients as Vitamin for Online Finetuning Decision Transformers

Add code
Oct 31, 2024
Viaarxiv icon

ReferEverything: Towards Segmenting Everything We Can Speak of in Videos

Add code
Oct 30, 2024
Figure 1 for ReferEverything: Towards Segmenting Everything We Can Speak of in Videos
Figure 2 for ReferEverything: Towards Segmenting Everything We Can Speak of in Videos
Figure 3 for ReferEverything: Towards Segmenting Everything We Can Speak of in Videos
Figure 4 for ReferEverything: Towards Segmenting Everything We Can Speak of in Videos
Viaarxiv icon

Swiss Army Knife: Synergizing Biases in Knowledge from Vision Foundation Models for Multi-Task Learning

Add code
Oct 18, 2024
Viaarxiv icon

SceneCraft: Layout-Guided 3D Scene Generation

Add code
Oct 11, 2024
Figure 1 for SceneCraft: Layout-Guided 3D Scene Generation
Figure 2 for SceneCraft: Layout-Guided 3D Scene Generation
Figure 3 for SceneCraft: Layout-Guided 3D Scene Generation
Figure 4 for SceneCraft: Layout-Guided 3D Scene Generation
Viaarxiv icon

Emerging Pixel Grounding in Large Multimodal Models Without Grounding Supervision

Add code
Oct 10, 2024
Viaarxiv icon