Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Unrolling SGD: Understanding Factors Influencing Machine Unlearning

About

Machine unlearning is the process through which a deployed machine learning model is made to forget about some of its training data points. While naively retraining the model from scratch is an option, it is almost always associated with large computational overheads for deep learning models. Thus, several approaches to approximately unlearn have been proposed along with corresponding metrics that formalize what it means for a model to forget about a data point. In this work, we first taxonomize approaches and metrics of approximate unlearning. As a result, we identify verification error, i.e., the L2 difference between the weights of an approximately unlearned and a naively retrained model, as an approximate unlearning metric that should be optimized for as it subsumes a large class of other metrics. We theoretically analyze the canonical training algorithm, stochastic gradient descent (SGD), to surface the variables which are relevant to reducing the verification error of approximate unlearning for SGD. From this analysis, we first derive an easy-to-compute proxy for verification error (termed unlearning error). The analysis also informs the design of a new training objective penalty that limits the overall change in weights during SGD and as a result facilitates approximate unlearning with lower verification error. We validate our theoretical work through an empirical evaluation on learning with CIFAR-10, CIFAR-100, and IMDB sentiment analysis.

Anvith Thudi, Gabriel Deza, Varun Chandrasekaran, Nicolas Papernot• 2021

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-10 (test)
Accuracy89.73
3381
Machine UnlearningTiny-ImageNet (train)
Removal Accuracy (Train)14.26
41
Class UnlearningCIFAR-10
Retain Accuracy90.25
39
Visual Question AnsweringCLEAR 1.0 (Retain)
Accuracy65.9
32
Single-class UnlearningCIFAR-100
ACCr72.05
28
Single-class UnlearningMNIST
Accuracy Retention (ACCr)0.9826
28
Machine UnlearningUnlearning Evaluation Dataset Forget, Retain, and Non-member sets
SQS79.8
24
Machine Unlearning5% (forget set)
GF154.7
24
Multimodal Machine UnlearningMLLMU-Bench LLaVA-1.5-7B (test 1)
Forget Rate60.9
24
Multimodal Machine UnlearningMLLMU-Bench LLaVA-1.5-7B (test 2)
Forget Rate61.3
24
Showing 10 of 76 rows
...

Other info

Follow for update