Picture for Kechun Xu

Kechun Xu

Grounding 3D Object Affordance with Language Instructions, Visual Observations and Interactions

Add code
Apr 07, 2025
Viaarxiv icon

Efficient Alignment of Unconditioned Action Prior for Language-conditioned Pick and Place in Clutter

Add code
Mar 12, 2025
Viaarxiv icon

Revisit Mixture Models for Multi-Agent Simulation: Experimental Study within a Unified Framework

Add code
Jan 28, 2025
Figure 1 for Revisit Mixture Models for Multi-Agent Simulation: Experimental Study within a Unified Framework
Figure 2 for Revisit Mixture Models for Multi-Agent Simulation: Experimental Study within a Unified Framework
Figure 3 for Revisit Mixture Models for Multi-Agent Simulation: Experimental Study within a Unified Framework
Figure 4 for Revisit Mixture Models for Multi-Agent Simulation: Experimental Study within a Unified Framework
Viaarxiv icon

Grasp, See and Place: Efficient Unknown Object Rearrangement with Policy Structure Prior

Add code
Feb 23, 2024
Viaarxiv icon

A Hyper-network Based End-to-end Visual Servoing with Arbitrary Desired Poses

Add code
Apr 18, 2023
Figure 1 for A Hyper-network Based End-to-end Visual Servoing with Arbitrary Desired Poses
Figure 2 for A Hyper-network Based End-to-end Visual Servoing with Arbitrary Desired Poses
Figure 3 for A Hyper-network Based End-to-end Visual Servoing with Arbitrary Desired Poses
Figure 4 for A Hyper-network Based End-to-end Visual Servoing with Arbitrary Desired Poses
Viaarxiv icon

Object-centric Inference for Language Conditioned Placement: A Foundation Model based Approach

Add code
Apr 06, 2023
Viaarxiv icon

Failure-aware Policy Learning for Self-assessable Robotics Tasks

Add code
Feb 25, 2023
Figure 1 for Failure-aware Policy Learning for Self-assessable Robotics Tasks
Figure 2 for Failure-aware Policy Learning for Self-assessable Robotics Tasks
Figure 3 for Failure-aware Policy Learning for Self-assessable Robotics Tasks
Figure 4 for Failure-aware Policy Learning for Self-assessable Robotics Tasks
Viaarxiv icon

A Joint Modeling of Vision-Language-Action for Target-oriented Grasping in Clutter

Add code
Feb 24, 2023
Figure 1 for A Joint Modeling of Vision-Language-Action for Target-oriented Grasping in Clutter
Figure 2 for A Joint Modeling of Vision-Language-Action for Target-oriented Grasping in Clutter
Figure 3 for A Joint Modeling of Vision-Language-Action for Target-oriented Grasping in Clutter
Figure 4 for A Joint Modeling of Vision-Language-Action for Target-oriented Grasping in Clutter
Viaarxiv icon

E-NeRV: Expedite Neural Video Representation with Disentangled Spatial-Temporal Context

Add code
Jul 17, 2022
Figure 1 for E-NeRV: Expedite Neural Video Representation with Disentangled Spatial-Temporal Context
Figure 2 for E-NeRV: Expedite Neural Video Representation with Disentangled Spatial-Temporal Context
Figure 3 for E-NeRV: Expedite Neural Video Representation with Disentangled Spatial-Temporal Context
Figure 4 for E-NeRV: Expedite Neural Video Representation with Disentangled Spatial-Temporal Context
Viaarxiv icon

Learning A Simulation-based Visual Policy for Real-world Peg In Unseen Holes

Add code
May 09, 2022
Figure 1 for Learning A Simulation-based Visual Policy for Real-world Peg In Unseen Holes
Figure 2 for Learning A Simulation-based Visual Policy for Real-world Peg In Unseen Holes
Figure 3 for Learning A Simulation-based Visual Policy for Real-world Peg In Unseen Holes
Figure 4 for Learning A Simulation-based Visual Policy for Real-world Peg In Unseen Holes
Viaarxiv icon