Picture for Xuanyao Chen

Xuanyao Chen

SparseViT: Revisiting Activation Sparsity for Efficient High-Resolution Vision Transformer

Add code
Mar 30, 2023
Viaarxiv icon

ViP3D: End-to-end Visual Trajectory Prediction via 3D Agent Queries

Add code
Aug 02, 2022
Figure 1 for ViP3D: End-to-end Visual Trajectory Prediction via 3D Agent Queries
Figure 2 for ViP3D: End-to-end Visual Trajectory Prediction via 3D Agent Queries
Figure 3 for ViP3D: End-to-end Visual Trajectory Prediction via 3D Agent Queries
Figure 4 for ViP3D: End-to-end Visual Trajectory Prediction via 3D Agent Queries
Viaarxiv icon

MUTR3D: A Multi-camera Tracking Framework via 3D-to-2D Queries

Add code
May 02, 2022
Figure 1 for MUTR3D: A Multi-camera Tracking Framework via 3D-to-2D Queries
Figure 2 for MUTR3D: A Multi-camera Tracking Framework via 3D-to-2D Queries
Figure 3 for MUTR3D: A Multi-camera Tracking Framework via 3D-to-2D Queries
Figure 4 for MUTR3D: A Multi-camera Tracking Framework via 3D-to-2D Queries
Viaarxiv icon

FUTR3D: A Unified Sensor Fusion Framework for 3D Detection

Add code
Mar 20, 2022
Figure 1 for FUTR3D: A Unified Sensor Fusion Framework for 3D Detection
Figure 2 for FUTR3D: A Unified Sensor Fusion Framework for 3D Detection
Figure 3 for FUTR3D: A Unified Sensor Fusion Framework for 3D Detection
Figure 4 for FUTR3D: A Unified Sensor Fusion Framework for 3D Detection
Viaarxiv icon

What Makes Multimodal Learning Better than Single (Provably)

Add code
Jun 08, 2021
Figure 1 for What Makes Multimodal Learning Better than Single (Provably)
Figure 2 for What Makes Multimodal Learning Better than Single (Provably)
Figure 3 for What Makes Multimodal Learning Better than Single (Provably)
Figure 4 for What Makes Multimodal Learning Better than Single (Provably)
Viaarxiv icon