Picture for Yinpeng Dong

Yinpeng Dong

Self-Memory Alignment: Mitigating Factual Hallucinations with Generalized Improvement

Add code
Feb 26, 2025
Viaarxiv icon

STAIR: Improving Safety Alignment with Introspective Reasoning

Add code
Feb 04, 2025
Figure 1 for STAIR: Improving Safety Alignment with Introspective Reasoning
Figure 2 for STAIR: Improving Safety Alignment with Introspective Reasoning
Figure 3 for STAIR: Improving Safety Alignment with Introspective Reasoning
Figure 4 for STAIR: Improving Safety Alignment with Introspective Reasoning
Viaarxiv icon

Towards the Worst-case Robustness of Large Language Models

Add code
Jan 31, 2025
Figure 1 for Towards the Worst-case Robustness of Large Language Models
Figure 2 for Towards the Worst-case Robustness of Large Language Models
Figure 3 for Towards the Worst-case Robustness of Large Language Models
Figure 4 for Towards the Worst-case Robustness of Large Language Models
Viaarxiv icon

Libra-Leaderboard: Towards Responsible AI through a Balanced Leaderboard of Safety and Capability

Add code
Dec 24, 2024
Figure 1 for Libra-Leaderboard: Towards Responsible AI through a Balanced Leaderboard of Safety and Capability
Figure 2 for Libra-Leaderboard: Towards Responsible AI through a Balanced Leaderboard of Safety and Capability
Figure 3 for Libra-Leaderboard: Towards Responsible AI through a Balanced Leaderboard of Safety and Capability
Figure 4 for Libra-Leaderboard: Towards Responsible AI through a Balanced Leaderboard of Safety and Capability
Viaarxiv icon

AdvDreamer Unveils: Are Vision-Language Models Truly Ready for Real-World 3D Variations?

Add code
Dec 04, 2024
Figure 1 for AdvDreamer Unveils: Are Vision-Language Models Truly Ready for Real-World 3D Variations?
Figure 2 for AdvDreamer Unveils: Are Vision-Language Models Truly Ready for Real-World 3D Variations?
Figure 3 for AdvDreamer Unveils: Are Vision-Language Models Truly Ready for Real-World 3D Variations?
Figure 4 for AdvDreamer Unveils: Are Vision-Language Models Truly Ready for Real-World 3D Variations?
Viaarxiv icon

Exploring Aleatoric Uncertainty in Object Detection via Vision Foundation Models

Add code
Nov 26, 2024
Figure 1 for Exploring Aleatoric Uncertainty in Object Detection via Vision Foundation Models
Figure 2 for Exploring Aleatoric Uncertainty in Object Detection via Vision Foundation Models
Figure 3 for Exploring Aleatoric Uncertainty in Object Detection via Vision Foundation Models
Figure 4 for Exploring Aleatoric Uncertainty in Object Detection via Vision Foundation Models
Viaarxiv icon

Scaling Laws for Black box Adversarial Attacks

Add code
Nov 25, 2024
Figure 1 for Scaling Laws for Black box Adversarial Attacks
Figure 2 for Scaling Laws for Black box Adversarial Attacks
Figure 3 for Scaling Laws for Black box Adversarial Attacks
Figure 4 for Scaling Laws for Black box Adversarial Attacks
Viaarxiv icon

Real-world Adversarial Defense against Patch Attacks based on Diffusion Model

Add code
Sep 14, 2024
Viaarxiv icon

T2VSafetyBench: Evaluating the Safety of Text-to-Video Generative Models

Add code
Jul 08, 2024
Viaarxiv icon

Benchmarking Trustworthiness of Multimodal Large Language Models: A Comprehensive Study

Add code
Jun 11, 2024
Figure 1 for Benchmarking Trustworthiness of Multimodal Large Language Models: A Comprehensive Study
Figure 2 for Benchmarking Trustworthiness of Multimodal Large Language Models: A Comprehensive Study
Figure 3 for Benchmarking Trustworthiness of Multimodal Large Language Models: A Comprehensive Study
Figure 4 for Benchmarking Trustworthiness of Multimodal Large Language Models: A Comprehensive Study
Viaarxiv icon