Picture for Wenqi Zhang

Wenqi Zhang

Think Twice, Click Once: Enhancing GUI Grounding via Fast and Slow Systems

Add code
Mar 09, 2025
Viaarxiv icon

DB-Explore: Automated Database Exploration and Instruction Synthesis for Text-to-SQL

Add code
Mar 06, 2025
Viaarxiv icon

AskToAct: Enhancing LLMs Tool Use via Self-Correcting Clarification

Add code
Mar 03, 2025
Viaarxiv icon

Efficient Semantic-aware Encryption for Secure Communications in Intelligent Connected Vehicles

Add code
Feb 23, 2025
Viaarxiv icon

STaR-SQL: Self-Taught Reasoner for Text-to-SQL

Add code
Feb 19, 2025
Viaarxiv icon

RG-Attn: Radian Glue Attention for Multi-modality Multi-agent Cooperative Perception

Add code
Jan 28, 2025
Viaarxiv icon

ECBench: Can Multi-modal Foundation Models Understand the Egocentric World? A Holistic Embodied Cognition Benchmark

Add code
Jan 09, 2025
Figure 1 for ECBench: Can Multi-modal Foundation Models Understand the Egocentric World? A Holistic Embodied Cognition Benchmark
Figure 2 for ECBench: Can Multi-modal Foundation Models Understand the Egocentric World? A Holistic Embodied Cognition Benchmark
Figure 3 for ECBench: Can Multi-modal Foundation Models Understand the Egocentric World? A Holistic Embodied Cognition Benchmark
Figure 4 for ECBench: Can Multi-modal Foundation Models Understand the Egocentric World? A Holistic Embodied Cognition Benchmark
Viaarxiv icon

2.5 Years in Class: A Multimodal Textbook for Vision-Language Pretraining

Add code
Jan 03, 2025
Figure 1 for 2.5 Years in Class: A Multimodal Textbook for Vision-Language Pretraining
Figure 2 for 2.5 Years in Class: A Multimodal Textbook for Vision-Language Pretraining
Figure 3 for 2.5 Years in Class: A Multimodal Textbook for Vision-Language Pretraining
Figure 4 for 2.5 Years in Class: A Multimodal Textbook for Vision-Language Pretraining
Viaarxiv icon

GaVaMoE: Gaussian-Variational Gated Mixture of Experts for Explainable Recommendation

Add code
Oct 15, 2024
Figure 1 for GaVaMoE: Gaussian-Variational Gated Mixture of Experts for Explainable Recommendation
Figure 2 for GaVaMoE: Gaussian-Variational Gated Mixture of Experts for Explainable Recommendation
Figure 3 for GaVaMoE: Gaussian-Variational Gated Mixture of Experts for Explainable Recommendation
Figure 4 for GaVaMoE: Gaussian-Variational Gated Mixture of Experts for Explainable Recommendation
Viaarxiv icon

Entering Real Social World! Benchmarking the Theory of Mind and Socialization Capabilities of LLMs from a First-person Perspective

Add code
Oct 08, 2024
Viaarxiv icon