Picture for Leo Yu Zhang

Leo Yu Zhang

Character-Level Perturbations Disrupt LLM Watermarks

Add code
Sep 11, 2025
Viaarxiv icon

Towards Reliable Forgetting: A Survey on Machine Unlearning Verification, Challenges, and Future Directions

Add code
Jun 18, 2025
Viaarxiv icon

TED-LaST: Towards Robust Backdoor Defense Against Adaptive Attacks

Add code
Jun 12, 2025
Viaarxiv icon

When Better Features Mean Greater Risks: The Performance-Privacy Trade-Off in Contrastive Learning

Add code
Jun 06, 2025
Viaarxiv icon

Performance Guaranteed Poisoning Attacks in Federated Learning: A Sliding Mode Approach

Add code
May 22, 2025
Viaarxiv icon

Secure Transfer Learning: Training Clean Models Against Backdoor in (Both) Pre-trained Encoders and Downstream Datasets

Add code
Apr 16, 2025
Viaarxiv icon

Exploring Gradient-Guided Masked Language Model to Detect Textual Adversarial Attacks

Add code
Apr 08, 2025
Viaarxiv icon

Test-Time Backdoor Detection for Object Detection Models

Add code
Mar 19, 2025
Viaarxiv icon

Improving Generalization of Universal Adversarial Perturbation via Dynamic Maximin Optimization

Add code
Mar 17, 2025
Viaarxiv icon

Not All Edges are Equally Robust: Evaluating the Robustness of Ranking-Based Federated Learning

Add code
Mar 12, 2025
Viaarxiv icon