Picture for Pengfei Xia

Pengfei Xia

Large Language Models are Good Attackers: Efficient and Stealthy Textual Backdoor Attacks

Add code
Aug 21, 2024
Viaarxiv icon

Efficient Trigger Word Insertion

Add code
Nov 23, 2023
Figure 1 for Efficient Trigger Word Insertion
Figure 2 for Efficient Trigger Word Insertion
Figure 3 for Efficient Trigger Word Insertion
Figure 4 for Efficient Trigger Word Insertion
Viaarxiv icon

Explore the Effect of Data Selection on Poison Efficiency in Backdoor Attacks

Add code
Oct 15, 2023
Figure 1 for Explore the Effect of Data Selection on Poison Efficiency in Backdoor Attacks
Figure 2 for Explore the Effect of Data Selection on Poison Efficiency in Backdoor Attacks
Figure 3 for Explore the Effect of Data Selection on Poison Efficiency in Backdoor Attacks
Figure 4 for Explore the Effect of Data Selection on Poison Efficiency in Backdoor Attacks
Viaarxiv icon

A Proxy-Free Strategy for Practically Improving the Poisoning Efficiency in Backdoor Attacks

Add code
Jun 14, 2023
Viaarxiv icon

Efficient Backdoor Attacks for Deep Neural Networks in Real-world Scenarios

Add code
Jun 14, 2023
Viaarxiv icon

Frequency Decomposition to Tap the Potential of Single Domain for Generalization

Add code
Apr 14, 2023
Viaarxiv icon

Data-Efficient Backdoor Attacks

Add code
Apr 22, 2022
Figure 1 for Data-Efficient Backdoor Attacks
Figure 2 for Data-Efficient Backdoor Attacks
Figure 3 for Data-Efficient Backdoor Attacks
Figure 4 for Data-Efficient Backdoor Attacks
Viaarxiv icon

A Statistical Difference Reduction Method for Escaping Backdoor Detection

Add code
Nov 09, 2021
Figure 1 for A Statistical Difference Reduction Method for Escaping Backdoor Detection
Figure 2 for A Statistical Difference Reduction Method for Escaping Backdoor Detection
Figure 3 for A Statistical Difference Reduction Method for Escaping Backdoor Detection
Figure 4 for A Statistical Difference Reduction Method for Escaping Backdoor Detection
Viaarxiv icon

Tightening the Approximation Error of Adversarial Risk with Auto Loss Function Search

Add code
Nov 09, 2021
Figure 1 for Tightening the Approximation Error of Adversarial Risk with Auto Loss Function Search
Figure 2 for Tightening the Approximation Error of Adversarial Risk with Auto Loss Function Search
Figure 3 for Tightening the Approximation Error of Adversarial Risk with Auto Loss Function Search
Figure 4 for Tightening the Approximation Error of Adversarial Risk with Auto Loss Function Search
Viaarxiv icon

Are High-Frequency Components Beneficial for Training of Generative Adversarial Networks

Add code
Mar 20, 2021
Figure 1 for Are High-Frequency Components Beneficial for Training of Generative Adversarial Networks
Figure 2 for Are High-Frequency Components Beneficial for Training of Generative Adversarial Networks
Figure 3 for Are High-Frequency Components Beneficial for Training of Generative Adversarial Networks
Figure 4 for Are High-Frequency Components Beneficial for Training of Generative Adversarial Networks
Viaarxiv icon