Picture for Will Cai

Will Cai

Michael Pokorny

Improving LLM Safety Alignment with Dual-Objective Optimization

Add code
Mar 05, 2025
Viaarxiv icon

Humanity's Last Exam

Add code
Jan 24, 2025
Viaarxiv icon

Scaling Laws for Data Poisoning in LLMs

Add code
Aug 06, 2024
Viaarxiv icon