Picture for Nicolas Papernot

Nicolas Papernot

Temporal-Difference Learning Using Distributed Error Signals

Add code
Nov 06, 2024
Viaarxiv icon

A False Sense of Safety: Unsafe Information Leakage in 'Safe' AI Responses

Add code
Jul 02, 2024
Viaarxiv icon

UnUnlearning: Unlearning is not sufficient for content regulation in advanced generative AI

Add code
Jun 27, 2024
Viaarxiv icon

LLM Dataset Inference: Did you train on my dataset?

Add code
Jun 10, 2024
Viaarxiv icon

Tighter Privacy Auditing of DP-SGD in the Hidden State Threat Model

Add code
May 23, 2024
Viaarxiv icon

Fairness Feedback Loops: Training on Synthetic Data Amplifies Bias

Add code
Mar 12, 2024
Viaarxiv icon

Inexact Unlearning Needs More Careful Evaluations to Avoid a False Sense of Privacy

Add code
Mar 02, 2024
Viaarxiv icon

Architectural Neural Backdoors from First Principles

Add code
Feb 10, 2024
Figure 1 for Architectural Neural Backdoors from First Principles
Figure 2 for Architectural Neural Backdoors from First Principles
Figure 3 for Architectural Neural Backdoors from First Principles
Figure 4 for Architectural Neural Backdoors from First Principles
Viaarxiv icon

Regulation Games for Trustworthy Machine Learning

Add code
Feb 05, 2024
Viaarxiv icon

Unlearnable Algorithms for In-context Learning

Add code
Feb 01, 2024
Viaarxiv icon