Picture for Dongxia Wang

Dongxia Wang

RAS-Eval: A Comprehensive Benchmark for Security Evaluation of LLM Agents in Real-World Environments

Add code
Jun 18, 2025
Viaarxiv icon

LightKG: Efficient Knowledge-Aware Recommendations with Simplified GNN Architecture

Add code
Jun 12, 2025
Viaarxiv icon

Fair-PP: A Synthetic Dataset for Aligning LLM with Personalized Preferences of Social Equity

Add code
May 17, 2025
Viaarxiv icon

Enhancing New-item Fairness in Dynamic Recommender Systems

Add code
Apr 30, 2025
Viaarxiv icon

A Comprehensive Survey in LLM(-Agent) Full Stack Safety: Data, Training and Deployment

Add code
Apr 22, 2025
Viaarxiv icon

TruePose: Human-Parsing-guided Attention Diffusion for Full-ID Preserving Pose Transfer

Add code
Feb 05, 2025
Viaarxiv icon

Defending LVLMs Against Vision Attacks through Partial-Perception Supervision

Add code
Dec 17, 2024
Figure 1 for Defending LVLMs Against Vision Attacks through Partial-Perception Supervision
Figure 2 for Defending LVLMs Against Vision Attacks through Partial-Perception Supervision
Figure 3 for Defending LVLMs Against Vision Attacks through Partial-Perception Supervision
Figure 4 for Defending LVLMs Against Vision Attacks through Partial-Perception Supervision
Viaarxiv icon

MRP-LLM: Multitask Reflective Large Language Models for Privacy-Preserving Next POI Recommendation

Add code
Dec 03, 2024
Viaarxiv icon

Characterizing and Evaluating the Reliability of LLMs against Jailbreak Attacks

Add code
Aug 18, 2024
Figure 1 for Characterizing and Evaluating the Reliability of LLMs against Jailbreak Attacks
Figure 2 for Characterizing and Evaluating the Reliability of LLMs against Jailbreak Attacks
Figure 3 for Characterizing and Evaluating the Reliability of LLMs against Jailbreak Attacks
Figure 4 for Characterizing and Evaluating the Reliability of LLMs against Jailbreak Attacks
Viaarxiv icon

S-Eval: Automatic and Adaptive Test Generation for Benchmarking Safety Evaluation of Large Language Models

Add code
May 28, 2024
Viaarxiv icon