Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Reflection Backdoor: A Natural Backdoor Attack on Deep Neural Networks

About

Recent studies have shown that DNNs can be compromised by backdoor attacks crafted at training time. A backdoor attack installs a backdoor into the victim model by injecting a backdoor pattern into a small proportion of the training data. At test time, the victim model behaves normally on clean test data, yet consistently predicts a specific (likely incorrect) target class whenever the backdoor pattern is present in a test example. While existing backdoor attacks are effective, they are not stealthy. The modifications made on training data or labels are often suspicious and can be easily detected by simple data filtering or human inspection. In this paper, we present a new type of backdoor attack inspired by an important natural phenomenon: reflection. Using mathematical modeling of physical reflection models, we propose reflection backdoor (Refool) to plant reflections as backdoor into a victim model. We demonstrate on 3 computer vision tasks and 5 datasets that, Refool can attack state-of-the-art DNNs with high success rate, and is resistant to state-of-the-art backdoor defenses.

Yunfei Liu, Xingjun Ma, James Bailey, Feng Lu• 2020

Related benchmarks

TaskDatasetResultRank
Image ClassificationGTSRB
Natural Accuracy95.1
87
Image ClassificationMNIST
Clean Accuracy97.7
71
Backdoor AttackImageNet 10-S
CAD2
13
Backdoor AttackCelebA-S
CAD12.36
13
Backdoor AttackCaltech-101
CAD-420
13
Backdoor AttackCars
CAD-5.12
13
Backdoor AttackPets
CAD-8.2
13
Backdoor AttackCIFAR-10-S
Clean Attack Drop (CAD)-0.67
13
Backdoor AttackBackdoor Attack Evaluation Summary
Poison Rate0.57
10
Image ClassificationCIFAR10
Clean Accuracy93.4
10
Showing 10 of 11 rows

Other info

Follow for update