Parameter Importance is Not Static: Evolving Parameter Isolation for Supervised Fine-Tuning
About
Supervised Fine-Tuning (SFT) of large language models often suffers from task interference and catastrophic forgetting. Recent approaches alleviate this issue by isolating task-critical parameters during training. However, these methods represent a static solution to a dynamic problem, assuming that parameter importance remains fixed once identified. In this work, we empirically demonstrate that parameter importance exhibits temporal drift over the course of training. To address this, we propose Evolving Parameter Isolation (EPI), a fine-tuning framework that adapts isolation decisions based on online estimates of parameter importance. Instead of freezing a fixed subset of parameters, EPI periodically updates isolation masks using gradient-based signals, enabling the model to protect emerging task-critical parameters while releasing outdated ones to recover plasticity. Experiments on diverse multi-task benchmarks demonstrate that EPI consistently reduces interference and forgetting compared to static isolation and standard fine-tuning, while improving overall generalization. Our analysis highlights the necessity of synchronizing isolation mechanisms with the evolving dynamics of learning diverse abilities.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Instruction Following | Alpaca | -- | 111 | |
| Supervised Fine-tuning | Heterogeneous SFT Benchmarks (Evaluation) | GSM8K Score65 | 20 | |
| Forgetting analysis | GSM8K, Code, LogiQA, Alpaca, UChat (test) | GSM8K Forgetting Rate4.8 | 15 | |
| Task Gradient Conflict | Sequential Fine-Tuning Suite (CodeAlpaca, LogiQA, Alpaca) (test) | Code Conflict Score13 | 15 | |
| Code Generation | CodeAlpaca | Score33.2 | 6 | |
| Conversational Instruction Following | UltraChat | Overall Score9.1 | 6 | |
| General Instruction Following Evaluation | Aggregated Multi-Task Suite | Average Normalized Score8.57 | 6 |