Picture for Ilia Shumailov

Ilia Shumailov

Defeating Prompt Injections by Design

Add code
Mar 24, 2025
Viaarxiv icon

Interpreting the Repeated Token Phenomenon in Large Language Models

Add code
Mar 11, 2025
Viaarxiv icon

Trusted Machine Learning Models Unlock Private Inference for Problems Currently Infeasible with Cryptography

Add code
Jan 15, 2025
Viaarxiv icon

Machine Unlearning Doesn't Do What You Think: Lessons for Generative AI Policy, Research, and Practice

Add code
Dec 09, 2024
Figure 1 for Machine Unlearning Doesn't Do What You Think: Lessons for Generative AI Policy, Research, and Practice
Figure 2 for Machine Unlearning Doesn't Do What You Think: Lessons for Generative AI Policy, Research, and Practice
Figure 3 for Machine Unlearning Doesn't Do What You Think: Lessons for Generative AI Policy, Research, and Practice
Figure 4 for Machine Unlearning Doesn't Do What You Think: Lessons for Generative AI Policy, Research, and Practice
Viaarxiv icon

Hardware and Software Platform Inference

Add code
Nov 07, 2024
Viaarxiv icon

Stealing User Prompts from Mixture of Experts

Add code
Oct 30, 2024
Viaarxiv icon

Measuring memorization through probabilistic discoverable extraction

Add code
Oct 25, 2024
Figure 1 for Measuring memorization through probabilistic discoverable extraction
Figure 2 for Measuring memorization through probabilistic discoverable extraction
Figure 3 for Measuring memorization through probabilistic discoverable extraction
Figure 4 for Measuring memorization through probabilistic discoverable extraction
Viaarxiv icon

Operationalizing Contextual Integrity in Privacy-Conscious Assistants

Add code
Aug 05, 2024
Figure 1 for Operationalizing Contextual Integrity in Privacy-Conscious Assistants
Figure 2 for Operationalizing Contextual Integrity in Privacy-Conscious Assistants
Figure 3 for Operationalizing Contextual Integrity in Privacy-Conscious Assistants
Figure 4 for Operationalizing Contextual Integrity in Privacy-Conscious Assistants
Viaarxiv icon

A False Sense of Safety: Unsafe Information Leakage in 'Safe' AI Responses

Add code
Jul 02, 2024
Viaarxiv icon

UnUnlearning: Unlearning is not sufficient for content regulation in advanced generative AI

Add code
Jun 27, 2024
Viaarxiv icon