Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Autoregressive Perturbations for Data Poisoning

About

The prevalence of data scraping from social media as a means to obtain datasets has led to growing concerns regarding unauthorized use of data. Data poisoning attacks have been proposed as a bulwark against scraping, as they make data "unlearnable" by adding small, imperceptible perturbations. Unfortunately, existing methods require knowledge of both the target architecture and the complete dataset so that a surrogate network can be trained, the parameters of which are used to generate the attack. In this work, we introduce autoregressive (AR) poisoning, a method that can generate poisoned data without access to the broader dataset. The proposed AR perturbations are generic, can be applied across different datasets, and can poison different architectures. Compared to existing unlearnable methods, our AR poisons are more resistant against common defenses such as adversarial training and strong data augmentations. Our analysis further provides insight into what makes an effective data poison.

Pedro Sandoval-Segura, Vasu Singla, Jonas Geiping, Micah Goldblum, Tom Goldstein, David W. Jacobs• 2022

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-10 (test)
Accuracy88.8
3381
Semantic segmentationADE20K (val)
mIoU43.9
2731
Semantic segmentationCityscapes (val)
mIoU68.9
287
Panoptic SegmentationCityscapes (val)
PQ51.6
276
Instance SegmentationCityscapes (val)
AP35.5
239
Panoptic SegmentationADE20K (val)
PQ37.8
89
Instance SegmentationADE20K (val)
AP25.4
21
Image ClassificationCIFAR-10 (test)
Accuracy (Base)16.89
11
Showing 8 of 8 rows

Other info

Follow for update