Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Eternal Sunshine of the Spotless Net: Selective Forgetting in Deep Networks

About

We explore the problem of selectively forgetting a particular subset of the data used for training a deep neural network. While the effects of the data to be forgotten can be hidden from the output of the network, insights may still be gleaned by probing deep into its weights. We propose a method for "scrubbing'" the weights clean of information about a particular set of training data. The method does not require retraining from scratch, nor access to the data originally used for training. Instead, the weights are modified so that any probing function of the weights is indistinguishable from the same function applied to the weights of a network trained without the data to be forgotten. This condition is a generalized and weaker form of Differential Privacy. Exploiting ideas related to the stability of stochastic gradient descent, we introduce an upper-bound on the amount of information remaining in the weights, which can be estimated efficiently even for deep neural networks.

Aditya Golatkar, Alessandro Achille, Stefano Soatto• 2019

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-10
Accuracy97.16
507
Semantic segmentationPascal VOC (test)--
236
Machine UnlearningCIFAR-10
Accf3.68
45
Machine UnlearningTiny-ImageNet (train)
Removal Accuracy (Train)99.98
41
Class UnlearningCIFAR-10
Retain Accuracy85.57
39
Selective UnlearningLacuna 10 (test)
Test Error (mean)1.53
36
Resolving ConfusionCIFAR-10
Test Error17.66
28
Machine UnlearningTiny-ImageNet Swin-T (test)
Residual Accuracy74.37
28
Single-class UnlearningCIFAR-100
ACCr70.97
28
Single-class UnlearningMNIST
Accuracy Retention (ACCr)0.9933
28
Showing 10 of 97 rows
...

Other info

Follow for update