Picture for Bingqian Lin

Bingqian Lin

PIVOT-R: Primitive-Driven Waypoint-Aware World Model for Robotic Manipulation

Add code
Oct 14, 2024
Figure 1 for PIVOT-R: Primitive-Driven Waypoint-Aware World Model for Robotic Manipulation
Figure 2 for PIVOT-R: Primitive-Driven Waypoint-Aware World Model for Robotic Manipulation
Figure 3 for PIVOT-R: Primitive-Driven Waypoint-Aware World Model for Robotic Manipulation
Figure 4 for PIVOT-R: Primitive-Driven Waypoint-Aware World Model for Robotic Manipulation
Viaarxiv icon

Affordances-Oriented Planning using Foundation Models for Continuous Vision-Language Navigation

Add code
Jul 08, 2024
Figure 1 for Affordances-Oriented Planning using Foundation Models for Continuous Vision-Language Navigation
Figure 2 for Affordances-Oriented Planning using Foundation Models for Continuous Vision-Language Navigation
Figure 3 for Affordances-Oriented Planning using Foundation Models for Continuous Vision-Language Navigation
Figure 4 for Affordances-Oriented Planning using Foundation Models for Continuous Vision-Language Navigation
Viaarxiv icon

Correctable Landmark Discovery via Large Models for Vision-Language Navigation

Add code
May 29, 2024
Figure 1 for Correctable Landmark Discovery via Large Models for Vision-Language Navigation
Figure 2 for Correctable Landmark Discovery via Large Models for Vision-Language Navigation
Figure 3 for Correctable Landmark Discovery via Large Models for Vision-Language Navigation
Figure 4 for Correctable Landmark Discovery via Large Models for Vision-Language Navigation
Viaarxiv icon

NavCoT: Boosting LLM-Based Vision-and-Language Navigation via Learning Disentangled Reasoning

Add code
Mar 12, 2024
Viaarxiv icon

Towards Deviation-Robust Agent Navigation via Perturbation-Aware Contrastive Learning

Add code
Mar 09, 2024
Viaarxiv icon

MapGPT: Map-Guided Prompting for Unified Vision-and-Language Navigation

Add code
Jan 14, 2024
Figure 1 for MapGPT: Map-Guided Prompting for Unified Vision-and-Language Navigation
Figure 2 for MapGPT: Map-Guided Prompting for Unified Vision-and-Language Navigation
Figure 3 for MapGPT: Map-Guided Prompting for Unified Vision-and-Language Navigation
Figure 4 for MapGPT: Map-Guided Prompting for Unified Vision-and-Language Navigation
Viaarxiv icon

Towards Medical Artificial General Intelligence via Knowledge-Enhanced Multimodal Pretraining

Add code
Apr 26, 2023
Figure 1 for Towards Medical Artificial General Intelligence via Knowledge-Enhanced Multimodal Pretraining
Figure 2 for Towards Medical Artificial General Intelligence via Knowledge-Enhanced Multimodal Pretraining
Figure 3 for Towards Medical Artificial General Intelligence via Knowledge-Enhanced Multimodal Pretraining
Figure 4 for Towards Medical Artificial General Intelligence via Knowledge-Enhanced Multimodal Pretraining
Viaarxiv icon

Dynamic Graph Enhanced Contrastive Learning for Chest X-ray Report Generation

Add code
Mar 18, 2023
Viaarxiv icon

Actional Atomic-Concept Learning for Demystifying Vision-Language Navigation

Add code
Feb 13, 2023
Viaarxiv icon

ADAPT: Vision-Language Navigation with Modality-Aligned Action Prompts

Add code
May 31, 2022
Figure 1 for ADAPT: Vision-Language Navigation with Modality-Aligned Action Prompts
Figure 2 for ADAPT: Vision-Language Navigation with Modality-Aligned Action Prompts
Figure 3 for ADAPT: Vision-Language Navigation with Modality-Aligned Action Prompts
Figure 4 for ADAPT: Vision-Language Navigation with Modality-Aligned Action Prompts
Viaarxiv icon