Picture for Xueyan Zou

Xueyan Zou

LoRA-TTT: Low-Rank Test-Time Training for Vision-Language Models

Add code
Feb 04, 2025
Viaarxiv icon

Integrating LMM Planners and 3D Skill Policies for Generalizable Manipulation

Add code
Jan 30, 2025
Viaarxiv icon

NaVILA: Legged Robot Vision-Language-Action Model for Navigation

Add code
Dec 05, 2024
Viaarxiv icon

WildLMa: Long Horizon Loco-Manipulation in the Wild

Add code
Nov 22, 2024
Figure 1 for WildLMa: Long Horizon Loco-Manipulation in the Wild
Figure 2 for WildLMa: Long Horizon Loco-Manipulation in the Wild
Figure 3 for WildLMa: Long Horizon Loco-Manipulation in the Wild
Figure 4 for WildLMa: Long Horizon Loco-Manipulation in the Wild
Viaarxiv icon

GraspSplats: Efficient Manipulation with 3D Feature Splatting

Add code
Sep 03, 2024
Viaarxiv icon

PLUM: Preference Learning Plus Test Cases Yields Better Code Language Models

Add code
Jun 11, 2024
Viaarxiv icon

Interfacing Foundation Models' Embeddings

Add code
Dec 12, 2023
Viaarxiv icon

LLaVA-Grounding: Grounded Visual Chat with Large Multimodal Models

Add code
Dec 05, 2023
Viaarxiv icon

Visual In-Context Prompting

Add code
Nov 22, 2023
Viaarxiv icon

LLaVA-Plus: Learning to Use Tools for Creating Multimodal Agents

Add code
Nov 09, 2023
Figure 1 for LLaVA-Plus: Learning to Use Tools for Creating Multimodal Agents
Figure 2 for LLaVA-Plus: Learning to Use Tools for Creating Multimodal Agents
Figure 3 for LLaVA-Plus: Learning to Use Tools for Creating Multimodal Agents
Figure 4 for LLaVA-Plus: Learning to Use Tools for Creating Multimodal Agents
Viaarxiv icon