Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Targeted Unlearning with Single Layer Unlearning Gradient

About

Machine unlearning methods aim to remove sensitive or unwanted content from trained models, but typically demand extensive model updates at significant computational cost while potentially degrading model performance on both related and unrelated tasks. We propose Single Layer Unlearning Gradient (SLUG) as an efficient method to unlearn targeted information by updating a single critical layer using a one-time gradient computation. SLUG uses layer importance and gradient alignment metrics to identify the optimal layer for targeted information removal while preserving the model utility. We demonstrate the effectiveness of SLUG for CLIP, Stable Diffusion, and vision-language models (VLMs) in removing concrete (e.g., identities and objects) and abstract concepts (e.g., artistic styles). On the UnlearnCanvas benchmark, SLUG achieves comparable unlearning performance to existing methods while requiring significantly less computational resources. Our proposed approach offers a practical solution for targeted unlearning that is computationally efficient and precise. Our code is available at https://github.com/CSIPlab/SLUG.

Zikui Cai, Yaoteng Tan, M. Salman Asif• 2024

Related benchmarks

TaskDatasetResultRank
Style UnlearningUnlearnCanvas
UA0.8629
25
Machine UnlearningImageNet
Forget Accuracy68.64
15
Machine UnlearningLAION 400M
Forget Accuracy48.46
15
Object UnlearningUnlearnCanvas
Unlearning Accuracy (UA)75.43
12
Machine UnlearningUnlearnCanvas SD-V1.5
Time (s)39
7
Object ForgettingUnlearnCanvas SD-V3
Unlearning Accuracy (UA)85.44
7
Object ForgettingUnlearnCanvas SD-V1.5
Unlearning Accuracy (UA)75.43
7
Style ForgettingUnlearnCanvas SD-V1.5
Unlearning Accuracy (UA)86.29
7
Style ForgettingUnlearnCanvas SD-V3
Unlearning Accuracy (UA)88.2
7
Showing 9 of 9 rows

Other info

Follow for update