Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Feature Squeezing: Detecting Adversarial Examples in Deep Neural Networks

About

Although deep neural networks (DNNs) have achieved great success in many tasks, they can often be fooled by \emph{adversarial examples} that are generated by adding small but purposeful distortions to natural examples. Previous studies to defend against adversarial examples mostly focused on refining the DNN models, but have either shown limited success or required expensive computation. We propose a new strategy, \emph{feature squeezing}, that can be used to harden DNN models by detecting adversarial examples. Feature squeezing reduces the search space available to an adversary by coalescing samples that correspond to many different feature vectors in the original space into a single sample. By comparing a DNN model's prediction on the original input with that on squeezed inputs, feature squeezing detects adversarial examples with high accuracy and few false positives. This paper explores two feature squeezing methods: reducing the color bit depth of each pixel and spatial smoothing. These simple strategies are inexpensive and complementary to other defenses, and can be combined in a joint detection framework to achieve high detection rates against state-of-the-art attacks.

Weilin Xu, David Evans, Yanjun Qi• 2017

Related benchmarks

TaskDatasetResultRank
Adversarial DetectionImageNet (val)
AUROC (PGD)94.71
14
Multi-task Driving Scene Understanding RobustnessThe Dolphins (Lvl. 0)
Final Score45.1
8
Multi-task Driving Scene Understanding RobustnessThe Dolphins Lvl. 1
Final Score45.59
8
Multi-task Driving Scene Understanding RobustnessThe Dolphins (Lvl. 2)
Final Score44.1
8
Multi-task Driving Scene Understanding RobustnessThe Dolphins (Lvl. 4)
Final Score40.6
8
Multi-task Driving Scene Understanding RobustnessThe Dolphins (Lvl. 3)
Final Score43.2
8
Showing 6 of 6 rows

Other info

Follow for update