Picture for Yingshui Tan

Yingshui Tan

BackdoorAgent: A Unified Framework for Backdoor Attacks on LLM-based Agents

Add code
Jan 08, 2026
Viaarxiv icon

Let It Flow: Agentic Crafting on Rock and Roll, Building the ROME Model within an Open Agentic Learning Ecosystem

Add code
Dec 31, 2025
Viaarxiv icon

QuadSentinel: Sequent Safety for Machine-Checkable Control in Multi-agent Systems

Add code
Dec 18, 2025
Figure 1 for QuadSentinel: Sequent Safety for Machine-Checkable Control in Multi-agent Systems
Figure 2 for QuadSentinel: Sequent Safety for Machine-Checkable Control in Multi-agent Systems
Figure 3 for QuadSentinel: Sequent Safety for Machine-Checkable Control in Multi-agent Systems
Figure 4 for QuadSentinel: Sequent Safety for Machine-Checkable Control in Multi-agent Systems
Viaarxiv icon

IFEvalCode: Controlled Code Generation

Add code
Jul 30, 2025
Viaarxiv icon

MSR-Align: Policy-Grounded Multimodal Alignment for Safety-Aware Reasoning in Vision-Language Models

Add code
Jun 24, 2025
Viaarxiv icon

Reinforcement Learning Optimization for Large-Scale Learning: An Efficient and User-Friendly Scaling Library

Add code
Jun 06, 2025
Viaarxiv icon

USB: A Comprehensive and Unified Safety Evaluation Benchmark for Multimodal Large Language Models

Add code
May 26, 2025
Figure 1 for USB: A Comprehensive and Unified Safety Evaluation Benchmark for Multimodal Large Language Models
Figure 2 for USB: A Comprehensive and Unified Safety Evaluation Benchmark for Multimodal Large Language Models
Figure 3 for USB: A Comprehensive and Unified Safety Evaluation Benchmark for Multimodal Large Language Models
Figure 4 for USB: A Comprehensive and Unified Safety Evaluation Benchmark for Multimodal Large Language Models
Viaarxiv icon

Beyond Safe Answers: A Benchmark for Evaluating True Risk Awareness in Large Reasoning Models

Add code
May 26, 2025
Figure 1 for Beyond Safe Answers: A Benchmark for Evaluating True Risk Awareness in Large Reasoning Models
Figure 2 for Beyond Safe Answers: A Benchmark for Evaluating True Risk Awareness in Large Reasoning Models
Figure 3 for Beyond Safe Answers: A Benchmark for Evaluating True Risk Awareness in Large Reasoning Models
Figure 4 for Beyond Safe Answers: A Benchmark for Evaluating True Risk Awareness in Large Reasoning Models
Viaarxiv icon

KORGym: A Dynamic Game Platform for LLM Reasoning Evaluation

Add code
May 21, 2025
Figure 1 for KORGym: A Dynamic Game Platform for LLM Reasoning Evaluation
Figure 2 for KORGym: A Dynamic Game Platform for LLM Reasoning Evaluation
Figure 3 for KORGym: A Dynamic Game Platform for LLM Reasoning Evaluation
Figure 4 for KORGym: A Dynamic Game Platform for LLM Reasoning Evaluation
Viaarxiv icon

DREAM: Disentangling Risks to Enhance Safety Alignment in Multimodal Large Language Models

Add code
Apr 25, 2025
Figure 1 for DREAM: Disentangling Risks to Enhance Safety Alignment in Multimodal Large Language Models
Figure 2 for DREAM: Disentangling Risks to Enhance Safety Alignment in Multimodal Large Language Models
Figure 3 for DREAM: Disentangling Risks to Enhance Safety Alignment in Multimodal Large Language Models
Figure 4 for DREAM: Disentangling Risks to Enhance Safety Alignment in Multimodal Large Language Models
Viaarxiv icon