Picture for Kamalika Chaudhuri

Kamalika Chaudhuri

UCSD

RL Is a Hammer and LLMs Are Nails: A Simple Reinforcement Learning Recipe for Strong Prompt Injection

Add code
Oct 06, 2025
Viaarxiv icon

Machine Learning with Privacy for Protected Attributes

Add code
Jun 24, 2025
Figure 1 for Machine Learning with Privacy for Protected Attributes
Figure 2 for Machine Learning with Privacy for Protected Attributes
Figure 3 for Machine Learning with Privacy for Protected Attributes
Figure 4 for Machine Learning with Privacy for Protected Attributes
Viaarxiv icon

Learning-Time Encoding Shapes Unlearning in LLMs

Add code
Jun 18, 2025
Viaarxiv icon

Can We Infer Confidential Properties of Training Data from LLMs?

Add code
Jun 12, 2025
Viaarxiv icon

AbstentionBench: Reasoning LLMs Fail on Unanswerable Questions

Add code
Jun 10, 2025
Viaarxiv icon

Do LLMs Really Forget? Evaluating Unlearning with Knowledge Correlation and Confidence Awareness

Add code
Jun 06, 2025
Viaarxiv icon

How much do language models memorize?

Add code
May 30, 2025
Viaarxiv icon

WASP: Benchmarking Web Agent Security Against Prompt Injection Attacks

Add code
Apr 30, 2025
Viaarxiv icon

Measuring Déjà vu Memorization Efficiently

Add code
Apr 08, 2025
Viaarxiv icon

AgentDAM: Privacy Leakage Evaluation for Autonomous Web Agents

Add code
Mar 12, 2025
Figure 1 for AgentDAM: Privacy Leakage Evaluation for Autonomous Web Agents
Figure 2 for AgentDAM: Privacy Leakage Evaluation for Autonomous Web Agents
Figure 3 for AgentDAM: Privacy Leakage Evaluation for Autonomous Web Agents
Figure 4 for AgentDAM: Privacy Leakage Evaluation for Autonomous Web Agents
Viaarxiv icon