Picture for Mark Dras

Mark Dras

Microsoft Research Institute, Macquarie University

Empirical Calibration and Metric Differential Privacy in Language Models

Add code
Mar 18, 2025
Viaarxiv icon

VaxGuard: A Multi-Generator, Multi-Type, and Multi-Role Dataset for Detecting LLM-Generated Vaccine Misinformation

Add code
Mar 12, 2025
Viaarxiv icon

VITAL: A New Dataset for Benchmarking Pluralistic Alignment in Healthcare

Add code
Feb 19, 2025
Viaarxiv icon

Comparing privacy notions for protection against reconstruction attacks in machine learning

Add code
Feb 06, 2025
Viaarxiv icon

Suspiciousness of Adversarial Texts to Human

Add code
Oct 06, 2024
Figure 1 for Suspiciousness of Adversarial Texts to Human
Figure 2 for Suspiciousness of Adversarial Texts to Human
Figure 3 for Suspiciousness of Adversarial Texts to Human
Figure 4 for Suspiciousness of Adversarial Texts to Human
Viaarxiv icon

IDT: Dual-Task Adversarial Attacks for Privacy Protection

Add code
Jun 28, 2024
Figure 1 for IDT: Dual-Task Adversarial Attacks for Privacy Protection
Figure 2 for IDT: Dual-Task Adversarial Attacks for Privacy Protection
Figure 3 for IDT: Dual-Task Adversarial Attacks for Privacy Protection
Figure 4 for IDT: Dual-Task Adversarial Attacks for Privacy Protection
Viaarxiv icon

Bayes' capacity as a measure for reconstruction attacks in federated learning

Add code
Jun 19, 2024
Figure 1 for Bayes' capacity as a measure for reconstruction attacks in federated learning
Figure 2 for Bayes' capacity as a measure for reconstruction attacks in federated learning
Viaarxiv icon

Seeing the Forest through the Trees: Data Leakage from Partial Transformer Gradients

Add code
Jun 03, 2024
Viaarxiv icon

Here's a Free Lunch: Sanitizing Backdoored Models with Model Merge

Add code
Feb 29, 2024
Viaarxiv icon

What Learned Representations and Influence Functions Can Tell Us About Adversarial Examples

Add code
Sep 21, 2023
Viaarxiv icon