Picture for Xiaoqi Li

Xiaoqi Li

ManipGPT: Is Affordance Segmentation by Large Vision Models Enough for Articulated Object Manipulation?

Add code
Dec 13, 2024
Viaarxiv icon

Human-centered In-building Embodied Delivery Benchmark

Add code
Jun 25, 2024
Viaarxiv icon

SpatialBot: Precise Spatial Understanding with Vision Language Models

Add code
Jun 19, 2024
Viaarxiv icon

AIC MLLM: Autonomous Interactive Correction MLLM for Robust Robotic Manipulation

Add code
Jun 17, 2024
Viaarxiv icon

GasTrace: Detecting Sandwich Attack Malicious Accounts in Ethereum

Add code
May 30, 2024
Viaarxiv icon

ManipVQA: Injecting Robotic Affordance and Physically Grounded Information into Multi-Modal Large Language Models

Add code
Mar 17, 2024
Viaarxiv icon

NaturalVLM: Leveraging Fine-grained Natural Language for Affordance-Guided Visual Manipulation

Add code
Mar 13, 2024
Viaarxiv icon

ManipLLM: Embodied Multimodal Large Language Model for Object-Centric Robotic Manipulation

Add code
Dec 24, 2023
Viaarxiv icon

LiDAR-LLM: Exploring the Potential of Large Language Models for 3D LiDAR Understanding

Add code
Dec 21, 2023
Viaarxiv icon

ImageManip: Image-based Robotic Manipulation with Affordance-guided Next View Selection

Add code
Oct 13, 2023
Viaarxiv icon