Picture for Nicolas Papernot

Nicolas Papernot

Pr$εε$mpt: Sanitizing Sensitive Prompts for LLMs

Add code
Apr 07, 2025
Viaarxiv icon

Backdoor Detection through Replicated Execution of Outsourced Training

Add code
Mar 31, 2025
Viaarxiv icon

Machine Unlearning Doesn't Do What You Think: Lessons for Generative AI Policy, Research, and Practice

Add code
Dec 09, 2024
Figure 1 for Machine Unlearning Doesn't Do What You Think: Lessons for Generative AI Policy, Research, and Practice
Figure 2 for Machine Unlearning Doesn't Do What You Think: Lessons for Generative AI Policy, Research, and Practice
Figure 3 for Machine Unlearning Doesn't Do What You Think: Lessons for Generative AI Policy, Research, and Practice
Figure 4 for Machine Unlearning Doesn't Do What You Think: Lessons for Generative AI Policy, Research, and Practice
Viaarxiv icon

On the Privacy Risk of In-context Learning

Add code
Nov 15, 2024
Viaarxiv icon

Temporal-Difference Learning Using Distributed Error Signals

Add code
Nov 06, 2024
Viaarxiv icon

A False Sense of Safety: Unsafe Information Leakage in 'Safe' AI Responses

Add code
Jul 02, 2024
Viaarxiv icon

UnUnlearning: Unlearning is not sufficient for content regulation in advanced generative AI

Add code
Jun 27, 2024
Viaarxiv icon

LLM Dataset Inference: Did you train on my dataset?

Add code
Jun 10, 2024
Figure 1 for LLM Dataset Inference: Did you train on my dataset?
Figure 2 for LLM Dataset Inference: Did you train on my dataset?
Figure 3 for LLM Dataset Inference: Did you train on my dataset?
Figure 4 for LLM Dataset Inference: Did you train on my dataset?
Viaarxiv icon

Tighter Privacy Auditing of DP-SGD in the Hidden State Threat Model

Add code
May 23, 2024
Viaarxiv icon

Fairness Feedback Loops: Training on Synthetic Data Amplifies Bias

Add code
Mar 12, 2024
Viaarxiv icon