Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Invisible Backdoor Attack against Self-supervised Learning

About

Self-supervised learning (SSL) models are vulnerable to backdoor attacks. Existing backdoor attacks that are effective in SSL often involve noticeable triggers, like colored patches or visible noise, which are vulnerable to human inspection. This paper proposes an imperceptible and effective backdoor attack against self-supervised models. We first find that existing imperceptible triggers designed for supervised learning are less effective in compromising self-supervised models. We then identify this ineffectiveness is attributed to the overlap in distributions between the backdoor and augmented samples used in SSL. Building on this insight, we design an attack using optimized triggers disentangled with the augmented transformation in the SSL, while remaining imperceptible to human vision. Experiments on five datasets and six SSL algorithms demonstrate our attack is highly effective and stealthy. It also has strong resistance to existing backdoor defenses. Our code can be found at https://github.com/Zhang-Henry/INACTIVE.

Hanrong Zhang, Zhenting Wang, Boheng Li, Fulin Lin, Tingxu Han, Mingyu Jin, Chenlu Zhan, Mengnan Du, Hongwei Wang, Shiqing Ma• 2024

Related benchmarks

TaskDatasetResultRank
Image ClassificationImageNet V2 (test)--
181
Image ClassificationImageNet-A (test)--
154
Image ClassificationImageNet-Sketch (test)--
132
Image ClassificationGTSRB--
87
Image-Text RetrievalCOCO (test)
Recall@139.46
37
Backdoor AttackCIFAR-10 (test)
Backdoor Accuracy93.01
30
Image ClassificationImageNet In-Distribution (test)
ID Accuracy58.62
23
Image ClassificationCIFAR-10
Clean Accuracy (CA)87.12
14
Image ClassificationCIFAR-100
Accuracy67.93
14
Image ClassificationSVHN
Accuracy (CA)59.25
14
Showing 10 of 28 rows

Other info

Code

Follow for update