Per-parameter Task Arithmetic for Unlearning in Large Language Models
About
In large language model (LLM) unlearning, private information is required to be removed. Task arithmetic unlearns by subtracting a specific task vector (TV)--defined as the parameter difference between a privacy-information-tuned model and the original model. While efficient, it can cause over-forgetting by disrupting parameters essential for retaining other information. Motivated by the observation that each parameter exhibits different importance for forgetting versus retention, we propose a per-parameter task arithmetic (PerTA) mechanism to rescale the TV, allowing per-parameter adjustment. These weights quantify the relative importance of each parameter for forgetting versus retention, estimated via gradients (i.e., PerTA-grad) or the diagonal Fisher information approximation (i.e., PerTA-fisher). Moreover, we discuss the effectiveness of PerTA, extend it to a more general form, and provide further analysis. Extensive experiments demonstrate that PerTA consistently improves upon standard TV, and in many cases surpasses widely used training-based unlearning methods in both forgetting effectiveness and overall model utility. By retaining the efficiency of task arithmetic while mitigating over-forgetting, PerTA offers a principled and practical framework for LLM unlearning.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| LLM Unlearning | TOFU 5% (forget set) | FQ-3.548 | 25 | |
| Machine Unlearning | TOFU 1% (forget set) | FQ-0.576 | 18 | |
| Machine Unlearning | TOFU 10% forget set 1.0 | FQ-2.99 | 18 | |
| Machine Unlearning | TOFU Llama-3.2 Instruct (average of 1%, 5%, 10% forget sets) | FQ Score-1.211 | 18 | |
| Machine Unlearning | MUSE | -- | 16 | |
| Knowledge Retention | MUSE Retain set (Dr) | KnowMem46.4 | 9 | |
| Machine Unlearning | TOFU 5% (forget set) | FQ-4.529 | 7 |