Picture for Aniruddha Saha

Aniruddha Saha

Generating Potent Poisons and Backdoors from Scratch with Guided Diffusion

Add code
Mar 25, 2024
Viaarxiv icon

Spotting LLMs With Binoculars: Zero-Shot Detection of Machine-Generated Text

Add code
Jan 22, 2024
Viaarxiv icon

NEFTune: Noisy Embeddings Improve Instruction Finetuning

Add code
Oct 10, 2023
Viaarxiv icon

Baseline Defenses for Adversarial Attacks Against Aligned Language Models

Add code
Sep 04, 2023
Viaarxiv icon

On the Reliability of Watermarks for Large Language Models

Add code
Jun 30, 2023
Viaarxiv icon

Bring Your Own Data! Self-Supervised Evaluation for Large Language Models

Add code
Jun 29, 2023
Viaarxiv icon

Revisiting Image Classifier Training for Improved Certified Robust Defense against Adversarial Patches

Add code
Jun 22, 2023
Viaarxiv icon

Backdoor Attacks on Vision Transformers

Add code
Jun 16, 2022
Figure 1 for Backdoor Attacks on Vision Transformers
Figure 2 for Backdoor Attacks on Vision Transformers
Figure 3 for Backdoor Attacks on Vision Transformers
Figure 4 for Backdoor Attacks on Vision Transformers
Viaarxiv icon

Backdoor Attacks on Self-Supervised Learning

Add code
May 21, 2021
Figure 1 for Backdoor Attacks on Self-Supervised Learning
Figure 2 for Backdoor Attacks on Self-Supervised Learning
Figure 3 for Backdoor Attacks on Self-Supervised Learning
Figure 4 for Backdoor Attacks on Self-Supervised Learning
Viaarxiv icon

Adversarial Patches Exploiting Contextual Reasoning in Object Detection

Add code
Sep 30, 2019
Figure 1 for Adversarial Patches Exploiting Contextual Reasoning in Object Detection
Figure 2 for Adversarial Patches Exploiting Contextual Reasoning in Object Detection
Figure 3 for Adversarial Patches Exploiting Contextual Reasoning in Object Detection
Figure 4 for Adversarial Patches Exploiting Contextual Reasoning in Object Detection
Viaarxiv icon