Picture for Gelei Deng

Gelei Deng

Risky-Bench: Probing Agentic Safety Risks under Real-World Deployment

Add code
Feb 03, 2026
Viaarxiv icon

DECEIVE-AFC: Adversarial Claim Attacks against Search-Enabled LLM-based Fact-Checking Systems

Add code
Jan 31, 2026
Viaarxiv icon

Self-Guard: Defending Large Reasoning Models via enhanced self-reflection

Add code
Jan 31, 2026
Viaarxiv icon

Agent Skills in the Wild: An Empirical Study of Security Vulnerabilities at Scale

Add code
Jan 15, 2026
Viaarxiv icon

Robust CAPTCHA Using Audio Illusions in the Era of Large Language Models: from Evaluation to Advances

Add code
Jan 13, 2026
Viaarxiv icon

PentestEval: Benchmarking LLM-based Penetration Testing with Modular and Stage-Level Design

Add code
Dec 16, 2025
Viaarxiv icon

RSafe: Incentivizing proactive reasoning to build robust and adaptive LLM safeguards

Add code
Jun 09, 2025
Figure 1 for RSafe: Incentivizing proactive reasoning to build robust and adaptive LLM safeguards
Figure 2 for RSafe: Incentivizing proactive reasoning to build robust and adaptive LLM safeguards
Figure 3 for RSafe: Incentivizing proactive reasoning to build robust and adaptive LLM safeguards
Figure 4 for RSafe: Incentivizing proactive reasoning to build robust and adaptive LLM safeguards
Viaarxiv icon

Holmes: Automated Fact Check with Large Language Models

Add code
May 06, 2025
Viaarxiv icon

A Rusty Link in the AI Supply Chain: Detecting Evil Configurations in Model Repositories

Add code
May 02, 2025
Figure 1 for A Rusty Link in the AI Supply Chain: Detecting Evil Configurations in Model Repositories
Figure 2 for A Rusty Link in the AI Supply Chain: Detecting Evil Configurations in Model Repositories
Viaarxiv icon

Enhancing Model Defense Against Jailbreaks with Proactive Safety Reasoning

Add code
Jan 31, 2025
Figure 1 for Enhancing Model Defense Against Jailbreaks with Proactive Safety Reasoning
Figure 2 for Enhancing Model Defense Against Jailbreaks with Proactive Safety Reasoning
Figure 3 for Enhancing Model Defense Against Jailbreaks with Proactive Safety Reasoning
Figure 4 for Enhancing Model Defense Against Jailbreaks with Proactive Safety Reasoning
Viaarxiv icon