Picture for Ziran Wang

Ziran Wang

SNM-Net: A Universal Framework for Robust Open-Set Gas Recognition via Spherical Normalization and Mahalanobis Distance

Add code
Dec 28, 2025
Viaarxiv icon

FSDAM: Few-Shot Driving Attention Modeling via Vision-Language Coupling

Add code
Nov 16, 2025
Figure 1 for FSDAM: Few-Shot Driving Attention Modeling via Vision-Language Coupling
Figure 2 for FSDAM: Few-Shot Driving Attention Modeling via Vision-Language Coupling
Figure 3 for FSDAM: Few-Shot Driving Attention Modeling via Vision-Language Coupling
Figure 4 for FSDAM: Few-Shot Driving Attention Modeling via Vision-Language Coupling
Viaarxiv icon

ViLaD: A Large Vision Language Diffusion Framework for End-to-End Autonomous Driving

Add code
Aug 18, 2025
Viaarxiv icon

A Hierarchical Test Platform for Vision Language Model (VLM)-Integrated Real-World Autonomous Driving

Add code
Jun 17, 2025
Viaarxiv icon

Inference Acceleration of Autoregressive Normalizing Flows by Selective Jacobi Decoding

Add code
May 30, 2025
Viaarxiv icon

ALN-P3: Unified Language Alignment for Perception, Prediction, and Planning in Autonomous Driving

Add code
May 21, 2025
Viaarxiv icon

Generative AI for Autonomous Driving: Frontiers and Opportunities

Add code
May 13, 2025
Viaarxiv icon

NuPlanQA: A Large-Scale Dataset and Benchmark for Multi-View Driving Scene Understanding in Multi-Modal Large Language Models

Add code
Mar 17, 2025
Figure 1 for NuPlanQA: A Large-Scale Dataset and Benchmark for Multi-View Driving Scene Understanding in Multi-Modal Large Language Models
Figure 2 for NuPlanQA: A Large-Scale Dataset and Benchmark for Multi-View Driving Scene Understanding in Multi-Modal Large Language Models
Figure 3 for NuPlanQA: A Large-Scale Dataset and Benchmark for Multi-View Driving Scene Understanding in Multi-Modal Large Language Models
Figure 4 for NuPlanQA: A Large-Scale Dataset and Benchmark for Multi-View Driving Scene Understanding in Multi-Modal Large Language Models
Viaarxiv icon

On-Board Vision-Language Models for Personalized Autonomous Vehicle Motion Control: System Design and Real-World Validation

Add code
Nov 17, 2024
Figure 1 for On-Board Vision-Language Models for Personalized Autonomous Vehicle Motion Control: System Design and Real-World Validation
Figure 2 for On-Board Vision-Language Models for Personalized Autonomous Vehicle Motion Control: System Design and Real-World Validation
Figure 3 for On-Board Vision-Language Models for Personalized Autonomous Vehicle Motion Control: System Design and Real-World Validation
Figure 4 for On-Board Vision-Language Models for Personalized Autonomous Vehicle Motion Control: System Design and Real-World Validation
Viaarxiv icon

MTA: Multimodal Task Alignment for BEV Perception and Captioning

Add code
Nov 16, 2024
Viaarxiv icon