Picture for Donglin Wang

Donglin Wang

Rethinking the Practicality of Vision-language-action Model: A Comprehensive Benchmark and An Improved Baseline

Add code
Feb 26, 2026
Viaarxiv icon

FRAPPE: Infusing World Modeling into Generalist Policies via Multiple Future Representation Alignment

Add code
Feb 19, 2026
Viaarxiv icon

CRL-VLA: Continual Vision-Language-Action Learning

Add code
Feb 03, 2026
Viaarxiv icon

CMR: Contractive Mapping Embeddings for Robust Humanoid Locomotion on Unstructured Terrains

Add code
Feb 03, 2026
Viaarxiv icon

Embodied Robot Manipulation in the Era of Foundation Models: Planning and Learning Perspectives

Add code
Dec 28, 2025
Viaarxiv icon

HiF-VLA: Hindsight, Insight and Foresight through Motion Representation for Vision-Language-Action Models

Add code
Dec 10, 2025
Figure 1 for HiF-VLA: Hindsight, Insight and Foresight through Motion Representation for Vision-Language-Action Models
Figure 2 for HiF-VLA: Hindsight, Insight and Foresight through Motion Representation for Vision-Language-Action Models
Figure 3 for HiF-VLA: Hindsight, Insight and Foresight through Motion Representation for Vision-Language-Action Models
Figure 4 for HiF-VLA: Hindsight, Insight and Foresight through Motion Representation for Vision-Language-Action Models
Viaarxiv icon

Executable Analytic Concepts as the Missing Link Between VLM Insight and Precise Manipulation

Add code
Oct 09, 2025
Viaarxiv icon

Robust Online Residual Refinement via Koopman-Guided Dynamics Modeling

Add code
Sep 16, 2025
Figure 1 for Robust Online Residual Refinement via Koopman-Guided Dynamics Modeling
Figure 2 for Robust Online Residual Refinement via Koopman-Guided Dynamics Modeling
Figure 3 for Robust Online Residual Refinement via Koopman-Guided Dynamics Modeling
Figure 4 for Robust Online Residual Refinement via Koopman-Guided Dynamics Modeling
Viaarxiv icon

Integrating Trajectory Optimization and Reinforcement Learning for Quadrupedal Jumping with Terrain-Adaptive Landing

Add code
Sep 16, 2025
Viaarxiv icon

VLA-Adapter: An Effective Paradigm for Tiny-Scale Vision-Language-Action Model

Add code
Sep 11, 2025
Viaarxiv icon