WISE: Rethinking the Knowledge Memory for Lifelong Model Editing of Large Language Models
About
Large language models (LLMs) need knowledge updates to meet the ever-growing world facts and correct the hallucinated responses, facilitating the methods of lifelong model editing. Where the updated knowledge resides in memories is a fundamental question for model editing. In this paper, we find that editing either long-term memory (direct model parameters) or working memory (non-parametric knowledge of neural network activations/representations by retrieval) will result in an impossible triangle -- reliability, generalization, and locality can not be realized together in the lifelong editing settings. For long-term memory, directly editing the parameters will cause conflicts with irrelevant pretrained knowledge or previous edits (poor reliability and locality). For working memory, retrieval-based activations can hardly make the model understand the edits and generalize (poor generalization). Therefore, we propose WISE to bridge the gap between memories. In WISE, we design a dual parametric memory scheme, which consists of the main memory for the pretrained knowledge and a side memory for the edited knowledge. We only edit the knowledge in the side memory and train a router to decide which memory to go through when given a query. For continual editing, we devise a knowledge-sharding mechanism where different sets of edits reside in distinct subspaces of parameters, and are subsequently merged into a shared memory without conflicts. Extensive experiments show that WISE can outperform previous model editing methods and overcome the impossible triangle under lifelong model editing of question answering, hallucination, and out-of-distribution settings across trending LLM architectures, e.g., GPT, LLaMA, and Mistral. Code is available at https://github.com/zjunlp/EasyEdit.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Knowledge Editing | zsRE | -- | 110 | |
| Knowledge Editing | ZsRE 10,000 facts | Reliability36.88 | 27 | |
| Knowledge Editing | Counterfact 10,000 facts | Relational Score1.84e+3 | 27 | |
| Knowledge Editing | E-VQA MMEdit 1.0 (test) | Reliability100 | 24 | |
| Knowledge Editing | MMEdit E-IC 1.0 (test) | Reliability100 | 24 | |
| Knowledge Editing | ZsRE (evaluation) | Reliability84 | 21 | |
| Knowledge Editing | Counterfact Full (test) | Rel. Accuracy74 | 21 | |
| Knowledge Editing | ZSRE (test) | Normalized Editing Time1.26 | 18 | |
| General Capability Preservation | SafeEdit | Fluency7.19 | 15 | |
| Lifelong Knowledge Editing | CounterFact | Reliability1.4 | 14 |