Attention Smoothing Is All You Need For Unlearning
About
Large Language Models are prone to memorizing sensitive, copyrighted, or hazardous content, posing significant privacy and legal concerns. Retraining from scratch is computationally infeasible, whereas current unlearning methods exhibit unstable trade-offs between forgetting and utility, frequently producing incoherent outputs on forget prompts and failing to generalize due to the persistence of lexical-level and semantic-level associations in attention. We propose Attention Smoothing Unlearning (ASU), a principled framework that casts unlearning as self-distillation from a forget-teacher derived from the model's own attention. By increasing the softmax temperature, ASU flattens attention distributions and directly suppresses the lexical-level and semantic-level associations responsible for reconstructing memorized knowledge. This results in a bounded optimization objective that erases factual information yet maintains coherence in responses to forget prompts. Empirical evaluation on TOFU, MUSE, and WMDP, along with real-world and continual unlearning scenarios across question answering and text completion, demonstrates that ASU outperforms the baselines for most unlearning scenarios, delivering robust unlearning with minimal loss of model utility.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Question Answering | MMLU | Accuracy57.5 | 46 | |
| Machine Unlearning | MUSE Books | Privacy Leakage-52.3 | 35 | |
| Machine Unlearning | TOFU 1.0 (forget01) | MU Score78.36 | 33 | |
| Machine Unlearning | TOFU forget10 1.0 | MU73.82 | 21 | |
| Machine Unlearning | TOFU forget05 1.0 | Model Utility (MU)74.18 | 21 | |
| Machine Unlearning | MUSE NEWS | VerbMem (Df)8.8 | 18 | |
| Personal Data Erasure | Synthetic Biographical Profiles | R-RL87 | 10 | |
| Misinformation Removal | MMLU corrupted ground truth | Accuracy (r)79.93 | 10 | |
| Harmful Knowledge Removal | WMDP bio | Acc_r68.39 | 10 | |
| Question Answering | WMDP multiple-choice QA | Bio Accuracy32.1 | 9 |