Picture for Jaden Mu

Jaden Mu

Making LLMs Vulnerable to Prompt Injection via Poisoning Alignment

Add code
Oct 18, 2024
Viaarxiv icon

An Investigation of Large Language Models for Real-World Hate Speech Detection

Add code
Jan 07, 2024
Viaarxiv icon