Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Per-parameter Task Arithmetic for Unlearning in Large Language Models

About

In large language model (LLM) unlearning, private information is required to be removed. Task arithmetic unlearns by subtracting a specific task vector (TV)--defined as the parameter difference between a privacy-information-tuned model and the original model. While efficient, it can cause over-forgetting by disrupting parameters essential for retaining other information. Motivated by the observation that each parameter exhibits different importance for forgetting versus retention, we propose a per-parameter task arithmetic (PerTA) mechanism to rescale the TV, allowing per-parameter adjustment. These weights quantify the relative importance of each parameter for forgetting versus retention, estimated via gradients (i.e., PerTA-grad) or the diagonal Fisher information approximation (i.e., PerTA-fisher). Moreover, we discuss the effectiveness of PerTA, extend it to a more general form, and provide further analysis. Extensive experiments demonstrate that PerTA consistently improves upon standard TV, and in many cases surpasses widely used training-based unlearning methods in both forgetting effectiveness and overall model utility. By retaining the efficiency of task arithmetic while mitigating over-forgetting, PerTA offers a principled and practical framework for LLM unlearning.

Chengyi Cai, Zesheng Ye, Jiangchao Yao, Jianzhong Qi, Bo Han, Xiaolu Zhang, Feng Liu, Jun Zhou• 2026

Related benchmarks

TaskDatasetResultRank
LLM UnlearningTOFU 5% (forget set)
FQ-3.548
25
Machine UnlearningTOFU 1% (forget set)
FQ-0.576
18
Machine UnlearningTOFU 10% forget set 1.0
FQ-2.99
18
Machine UnlearningTOFU Llama-3.2 Instruct (average of 1%, 5%, 10% forget sets)
FQ Score-1.211
18
Machine UnlearningMUSE--
16
Knowledge RetentionMUSE Retain set (Dr)
KnowMem46.4
9
Machine UnlearningTOFU 5% (forget set)
FQ-4.529
7
Showing 7 of 7 rows

Other info

Follow for update