Picture for Martin Pawelczyk

Martin Pawelczyk

Generalizing Trust: Weak-to-Strong Trustworthiness in Language Models

Add code
Dec 31, 2024
Viaarxiv icon

Explaining the Model, Protecting Your Data: Revealing and Mitigating the Data Privacy Risks of Post-Hoc Model Explanations via Membership Inference

Add code
Jul 24, 2024
Viaarxiv icon

Machine Unlearning Fails to Remove Data Poisoning Attacks

Add code
Jun 25, 2024
Viaarxiv icon

Towards Non-Adversarial Algorithmic Recourse

Add code
Mar 15, 2024
Viaarxiv icon

In-Context Unlearning: Language Models as Few Shot Unlearners

Add code
Oct 12, 2023
Viaarxiv icon

Gaussian Membership Inference Privacy

Add code
Jun 12, 2023
Viaarxiv icon

On the Privacy Risks of Algorithmic Recourse

Add code
Nov 10, 2022
Figure 1 for On the Privacy Risks of Algorithmic Recourse
Figure 2 for On the Privacy Risks of Algorithmic Recourse
Figure 3 for On the Privacy Risks of Algorithmic Recourse
Figure 4 for On the Privacy Risks of Algorithmic Recourse
Viaarxiv icon

Decomposing Counterfactual Explanations for Consequential Decision Making

Add code
Nov 03, 2022
Viaarxiv icon

I Prefer not to Say: Operationalizing Fair and User-guided Data Minimization

Add code
Nov 01, 2022
Viaarxiv icon

Language Models are Realistic Tabular Data Generators

Add code
Oct 12, 2022
Figure 1 for Language Models are Realistic Tabular Data Generators
Figure 2 for Language Models are Realistic Tabular Data Generators
Figure 3 for Language Models are Realistic Tabular Data Generators
Figure 4 for Language Models are Realistic Tabular Data Generators
Viaarxiv icon