Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Fine-tuning Done Right in Model Editing

About

Fine-tuning, a foundational method for adapting large language models, has long been considered ineffective for model editing. Here, we challenge this belief, arguing that the reported failure arises not from the inherent limitation of fine-tuning itself, but from adapting it to the sequential nature of the editing task, a single-pass depth-first pipeline that optimizes each sample to convergence before moving on. While intuitive, this depth-first pipeline coupled with sample-wise updating over-optimizes each edit and induces interference across edits. Our controlled experiments reveal that simply restoring fine-tuning to the standard breadth-first (i.e., epoch-based) pipeline with mini-batch optimization substantially improves its effectiveness for model editing. Moreover, fine-tuning in editing also suffers from suboptimal tuning parameter locations inherited from prior methods. Through systematic analysis of tuning locations, we derive LocFT-BF, a simple and effective localized editing method built on the restored fine-tuning framework. Extensive experiments across diverse LLMs and datasets demonstrate that LocFT-BF outperforms state-of-the-art methods by large margins. Notably, to our knowledge, it is the first to sustain 100K edits and 72B-parameter models,10 x beyond prior practice, without sacrificing general capabilities. By clarifying a long-standing misconception and introducing a principled localized tuning strategy, we advance fine-tuning from an underestimated baseline to a leading method for model editing, establishing a solid foundation for future research.

Wanli Yang, Rui Tang, Hongyu Zang, Du Su, Qi Cao, Jingang Wang, Huawei Shen, Xueqi Cheng, Fei Sun• 2025

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningGSM8K
Math Score73
171
Commonsense ReasoningARC-C
Accuracy50
51
Model EditingWikiBigEdit
MMLU69.2
34
Model EditingCounterFact
Reliability61.1
26
Model EditingzsRE
Reliability69.5
26
Model EditingzsRE
Reliability0.535
16
Multi-task Language UnderstandingMMLU
MMLU Score68
14
Model EditingZsRE 3,000 samples (test)
Relational Score99.1
13
Model EditingWikiBigEdit 3,000 samples (test)
Reliability99.9
13
Model EditingCounterFact 3,000 samples (test)
Reliability9.97e+3
13
Showing 10 of 15 rows

Other info

Follow for update