Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

RAP: Robustness-Aware Perturbations for Defending against Backdoor Attacks on NLP Models

About

Backdoor attacks, which maliciously control a well-trained model's outputs of the instances with specific triggers, are recently shown to be serious threats to the safety of reusing deep neural networks (DNNs). In this work, we propose an efficient online defense mechanism based on robustness-aware perturbations. Specifically, by analyzing the backdoor training process, we point out that there exists a big gap of robustness between poisoned and clean samples. Motivated by this observation, we construct a word-based robustness-aware perturbation to distinguish poisoned samples from clean samples to defend against the backdoor attacks on natural language processing (NLP) models. Moreover, we give a theoretical analysis about the feasibility of our robustness-aware perturbation-based defense method. Experimental results on sentiment analysis and toxic detection tasks show that our method achieves better defending performance and much lower computational costs than existing online defense methods. Our code is available at https://github.com/lancopku/RAP.

Wenkai Yang, Yankai Lin, Peng Li, Jie Zhou, Xu Sun• 2021

Related benchmarks

TaskDatasetResultRank
Backdoor DefenseAGNews
Attack Success Rate59.67
105
Poisoned sample detectionTrojAI round 6 (test)
Precision0.853
96
Sentiment ClassificationSST-2 64 instances (test)
Accuracy90.37
80
Backdoor DefenseAverage of four datasets (test)
Accuracy89.95
76
Topic ClassificationAG's News
ASR33.67
70
Backdoor DefenseSST-2
CACC91.71
65
Bias DefenseAverage of four datasets (test)
Accuracy89.98
56
Backdoor Attack ClassificationHSOL
ASR100
50
Sentiment AnalysisSST-2 (test)
Clean Accuracy91.93
50
Text ClassificationSubj
CA (%)0.967
48
Showing 10 of 31 rows

Other info

Follow for update