Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

UltraEdit: Training-, Subject-, and Memory-Free Lifelong Editing in Language Models

About

Lifelong learning enables large language models (LLMs) to adapt to evolving information by continually updating their internal knowledge. An ideal system should support efficient, wide-ranging updates while preserving existing capabilities and ensuring reliable deployment. Model editing stands out as a promising solution for this goal, offering a focused and efficient way to revise a model's internal knowledge. Although recent paradigms have made notable progress, they often struggle to meet the demands of practical lifelong adaptation at scale. To bridge this gap, we propose UltraEdit, a training-, subject-, and memory-free approach that is well-suited for ultra-scalable, real-world lifelong model editing. UltraEdit fundamentally differs from traditional paradigms by computing parameter shifts in one step using only a hidden state and its gradient, making the approach simple yet efficient. To improve scalability in lifelong settings, UltraEdit employs a lifelong normalization strategy that continuously updates feature statistics across turns, allowing it to adapt to distributional shifts and maintain consistency over time. UltraEdit achieves editing speeds more than $7\times$ faster than the previous state-of-the-art method, while requiring $4\times$ less VRAM. This makes it the only method currently capable of editing a 7B LLM on a 24GB consumer-grade GPU. Furthermore, we construct UltraEditBench, the largest dataset in the field to date with over 2M editing pairs, and demonstrate that our method supports up to 2M edits while maintaining high accuracy. Comprehensive experiments on five datasets and six models show that UltraEdit consistently achieves superior performance across diverse model editing scenarios, taking a further step towards safe and scalable lifelong learning. Our code is available at https://github.com/XiaojieGu/UltraEdit.

Xiaojie Gu, Ziying Huang, Jia-Chen Gu, Kai Zhang• 2025

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningGSM8K
Math Score73
197
Knowledge EditingzsRE
Generality77.08
181
Commonsense ReasoningARC-C
Accuracy46
172
Model EditingzsRE
Efficacy90.07
71
Model EditingUltraEditBench
Efficacy85.7
51
Model EditingFEVER
Efficacy98.23
49
Model EditingWikiBigEdit
Efficacy79.6
49
Model EditingWikiBigEdit
MMLU69.3
34
Model EditingzsRE
Reliability22.7
26
Model EditingCounterFact
Reliability18.1
26
Showing 10 of 22 rows

Other info

Follow for update