Picture for Edoardo Debenedetti

Edoardo Debenedetti

Design Patterns for Securing LLM Agents against Prompt Injections

Add code
Jun 11, 2025
Figure 1 for Design Patterns for Securing LLM Agents against Prompt Injections
Figure 2 for Design Patterns for Securing LLM Agents against Prompt Injections
Figure 3 for Design Patterns for Securing LLM Agents against Prompt Injections
Figure 4 for Design Patterns for Securing LLM Agents against Prompt Injections
Viaarxiv icon

LLMs unlock new paths to monetizing exploits

Add code
May 16, 2025
Figure 1 for LLMs unlock new paths to monetizing exploits
Figure 2 for LLMs unlock new paths to monetizing exploits
Figure 3 for LLMs unlock new paths to monetizing exploits
Figure 4 for LLMs unlock new paths to monetizing exploits
Viaarxiv icon

Defeating Prompt Injections by Design

Add code
Mar 24, 2025
Viaarxiv icon

AutoAdvExBench: Benchmarking autonomous exploitation of adversarial example defenses

Add code
Mar 03, 2025
Figure 1 for AutoAdvExBench: Benchmarking autonomous exploitation of adversarial example defenses
Figure 2 for AutoAdvExBench: Benchmarking autonomous exploitation of adversarial example defenses
Figure 3 for AutoAdvExBench: Benchmarking autonomous exploitation of adversarial example defenses
Figure 4 for AutoAdvExBench: Benchmarking autonomous exploitation of adversarial example defenses
Viaarxiv icon

Exploring Memorization and Copyright Violation in Frontier LLMs: A Study of the New York Times v. OpenAI 2023 Lawsuit

Add code
Dec 09, 2024
Figure 1 for Exploring Memorization and Copyright Violation in Frontier LLMs: A Study of the New York Times v. OpenAI 2023 Lawsuit
Figure 2 for Exploring Memorization and Copyright Violation in Frontier LLMs: A Study of the New York Times v. OpenAI 2023 Lawsuit
Figure 3 for Exploring Memorization and Copyright Violation in Frontier LLMs: A Study of the New York Times v. OpenAI 2023 Lawsuit
Figure 4 for Exploring Memorization and Copyright Violation in Frontier LLMs: A Study of the New York Times v. OpenAI 2023 Lawsuit
Viaarxiv icon

Measuring Non-Adversarial Reproduction of Training Data in Large Language Models

Add code
Nov 15, 2024
Viaarxiv icon

Adversarial Search Engine Optimization for Large Language Models

Add code
Jun 26, 2024
Figure 1 for Adversarial Search Engine Optimization for Large Language Models
Figure 2 for Adversarial Search Engine Optimization for Large Language Models
Figure 3 for Adversarial Search Engine Optimization for Large Language Models
Figure 4 for Adversarial Search Engine Optimization for Large Language Models
Viaarxiv icon

AgentDojo: A Dynamic Environment to Evaluate Attacks and Defenses for LLM Agents

Add code
Jun 19, 2024
Viaarxiv icon

Dataset and Lessons Learned from the 2024 SaTML LLM Capture-the-Flag Competition

Add code
Jun 12, 2024
Figure 1 for Dataset and Lessons Learned from the 2024 SaTML LLM Capture-the-Flag Competition
Figure 2 for Dataset and Lessons Learned from the 2024 SaTML LLM Capture-the-Flag Competition
Figure 3 for Dataset and Lessons Learned from the 2024 SaTML LLM Capture-the-Flag Competition
Figure 4 for Dataset and Lessons Learned from the 2024 SaTML LLM Capture-the-Flag Competition
Viaarxiv icon

AI Risk Management Should Incorporate Both Safety and Security

Add code
May 29, 2024
Figure 1 for AI Risk Management Should Incorporate Both Safety and Security
Viaarxiv icon