Picture for Shiguang Shan

Shiguang Shan

What Makes VLMs Robust? Towards Reconciling Robustness and Accuracy in Vision-Language Models

Add code
Mar 13, 2026
Viaarxiv icon

Neural Gate: Mitigating Privacy Risks in LVLMs via Neuron-Level Gradient Gating

Add code
Mar 13, 2026
Viaarxiv icon

INFACT: A Diagnostic Benchmark for Induced Faithfulness and Factuality Hallucinations in Video-LLMs

Add code
Mar 12, 2026
Viaarxiv icon

OSI: One-step Inversion Excels in Extracting Diffusion Watermarks

Add code
Feb 10, 2026
Viaarxiv icon

Contrastive Spectral Rectification: Test-Time Defense towards Zero-shot Adversarial Robustness of CLIP

Add code
Jan 27, 2026
Viaarxiv icon

CLIP-Guided Adaptable Self-Supervised Learning for Human-Centric Visual Tasks

Add code
Jan 19, 2026
Viaarxiv icon

T2VAttack: Adversarial Attack on Text-to-Video Diffusion Models

Add code
Dec 30, 2025
Viaarxiv icon

Steering Vision-Language Pre-trained Models for Incremental Face Presentation Attack Detection

Add code
Dec 24, 2025
Viaarxiv icon

Semantic Mismatch and Perceptual Degradation: A New Perspective on Image Editing Immunity

Add code
Dec 16, 2025
Figure 1 for Semantic Mismatch and Perceptual Degradation: A New Perspective on Image Editing Immunity
Figure 2 for Semantic Mismatch and Perceptual Degradation: A New Perspective on Image Editing Immunity
Figure 3 for Semantic Mismatch and Perceptual Degradation: A New Perspective on Image Editing Immunity
Figure 4 for Semantic Mismatch and Perceptual Degradation: A New Perspective on Image Editing Immunity
Viaarxiv icon

Dual Attention Guided Defense Against Malicious Edits

Add code
Dec 16, 2025
Viaarxiv icon