Picture for Nicolas Papernot

Nicolas Papernot

Machine Unlearning Doesn't Do What You Think: Lessons for Generative AI Policy, Research, and Practice

Add code
Dec 09, 2024
Viaarxiv icon

On the Privacy Risk of In-context Learning

Add code
Nov 15, 2024
Viaarxiv icon

Temporal-Difference Learning Using Distributed Error Signals

Add code
Nov 06, 2024
Viaarxiv icon

A False Sense of Safety: Unsafe Information Leakage in 'Safe' AI Responses

Add code
Jul 02, 2024
Viaarxiv icon

UnUnlearning: Unlearning is not sufficient for content regulation in advanced generative AI

Add code
Jun 27, 2024
Viaarxiv icon

LLM Dataset Inference: Did you train on my dataset?

Add code
Jun 10, 2024
Figure 1 for LLM Dataset Inference: Did you train on my dataset?
Figure 2 for LLM Dataset Inference: Did you train on my dataset?
Figure 3 for LLM Dataset Inference: Did you train on my dataset?
Figure 4 for LLM Dataset Inference: Did you train on my dataset?
Viaarxiv icon

Tighter Privacy Auditing of DP-SGD in the Hidden State Threat Model

Add code
May 23, 2024
Viaarxiv icon

Fairness Feedback Loops: Training on Synthetic Data Amplifies Bias

Add code
Mar 12, 2024
Viaarxiv icon

Inexact Unlearning Needs More Careful Evaluations to Avoid a False Sense of Privacy

Add code
Mar 02, 2024
Viaarxiv icon

Architectural Neural Backdoors from First Principles

Add code
Feb 10, 2024
Figure 1 for Architectural Neural Backdoors from First Principles
Figure 2 for Architectural Neural Backdoors from First Principles
Figure 3 for Architectural Neural Backdoors from First Principles
Figure 4 for Architectural Neural Backdoors from First Principles
Viaarxiv icon