Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Unified Gradient-Based Machine Unlearning with Remain Geometry Enhancement

About

Machine unlearning (MU) has emerged to enhance the privacy and trustworthiness of deep neural networks. Approximate MU is a practical method for large-scale models. Our investigation into approximate MU starts with identifying the steepest descent direction, minimizing the output Kullback-Leibler divergence to exact MU inside a parameters' neighborhood. This probed direction decomposes into three components: weighted forgetting gradient ascent, fine-tuning retaining gradient descent, and a weight saliency matrix. Such decomposition derived from Euclidean metric encompasses most existing gradient-based MU methods. Nevertheless, adhering to Euclidean space may result in sub-optimal iterative trajectories due to the overlooked geometric structure of the output probability space. We suggest embedding the unlearning update into a manifold rendered by the remaining geometry, incorporating second-order Hessian from the remaining data. It helps prevent effective unlearning from interfering with the retained performance. However, computing the second-order Hessian for large-scale models is intractable. To efficiently leverage the benefits of Hessian modulation, we propose a fast-slow parameter update strategy to implicitly approximate the up-to-date salient unlearning direction. Free from specific modal constraints, our approach is adaptable across computer vision unlearning tasks, including classification and generation. Extensive experiments validate our efficacy and efficiency. Notably, our method successfully performs class-forgetting on ImageNet using DiT and forgets a class on CIFAR-10 using DDPM in just 50 steps, compared to thousands of steps required by previous methods.

Zhehao Huang, Xinwen Cheng, JingHao Zheng, Haoran Wang, Zhengbao He, Tao Li, Xiaolin Huang• 2024

Related benchmarks

TaskDatasetResultRank
Machine UnlearningMUSE-Books Harry Potter v1.0 (Overall)
R-Forget10.45
17
Black-box Membership Inference AttackCIFAR-10 Most-memorized 1% forget samples
AUC0.518
12
Membership Inference AttackCIFAR-10 (Forget)
AUC50.9
12
Machine UnlearningCIFAR-100 Random Forget (10%)
FA62.69
11
Machine UnlearningTiny-Imagenet Random Forget 20%, γ=0 (test)
FA50.23
11
Machine UnlearningTiny-Imagenet Random Forget 30%, γ=1/3 (test)
FA51.2
11
Machine UnlearningCIFAR-100 Random Forget (forget set 20%)
FA65.3
11
Machine UnlearningTiny-Imagenet Random Forget 50%, γ=0 (test)
FA38.89
11
Machine UnlearningCIFAR-100 Random Forget (40%)
FA51.33
11
Machine UnlearningTiny-Imagenet Random Forget 10%, γ=1 (test)
FA22.01
11
Showing 10 of 14 rows

Other info

Follow for update