Picture for Pankayaraj Pathmanathan

Pankayaraj Pathmanathan

AdvBDGen: Adversarially Fortified Prompt-Specific Fuzzy Backdoor Generator Against LLM Alignment

Add code
Oct 15, 2024
Figure 1 for AdvBDGen: Adversarially Fortified Prompt-Specific Fuzzy Backdoor Generator Against LLM Alignment
Figure 2 for AdvBDGen: Adversarially Fortified Prompt-Specific Fuzzy Backdoor Generator Against LLM Alignment
Figure 3 for AdvBDGen: Adversarially Fortified Prompt-Specific Fuzzy Backdoor Generator Against LLM Alignment
Figure 4 for AdvBDGen: Adversarially Fortified Prompt-Specific Fuzzy Backdoor Generator Against LLM Alignment
Viaarxiv icon

Can Watermarking Large Language Models Prevent Copyrighted Text Generation and Hide Training Data?

Add code
Jul 24, 2024
Viaarxiv icon

Is poisoning a real threat to LLM alignment? Maybe more so than you think

Add code
Jun 17, 2024
Viaarxiv icon

Using Curiosity for an Even Representation of Tasks in Continual Offline Reinforcement Learning

Add code
Dec 05, 2023
Viaarxiv icon