EtCon: Edit-then-Consolidate for Reliable Knowledge Editing
About
Knowledge editing aims to update specific facts in large language models (LLMs) without full retraining. Prior efforts sought to tune the knowledge layers of LLMs, achieving improved performance in controlled, teacher-forced evaluations. However, they still encounter challenges in real-world autoregressive generation scenarios, which greatly limit their practical applicability. Our empirical analysis reveals two issues: (1) Most methods degrade pre-trained capabilities after injecting new knowledge; (2) They may exhibit a discrepancy between stored parametric knowledge and inference-time autoregressive generation behavior. To this end, we propose EtCon, an edit-then-consolidate paradigm that couples targeted edits with post-edit consolidation. Specifically, our framework comprises two stages: (1) Targeted Proximal Supervised Fine-Tuning (TPSFT) performs a constrained targeted edit to update parametric knowledge while controlling policy drift. (2) Group Relative Policy Optimization (GRPO) consolidates the edit by aligning autoregressive trajectories with the intended fact. Extensive experiments demonstrate that our EtCon improves editing reliability and real-world generalization, while better preserving pre-trained capabilities.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Question Answering | SQuAD | F119.6 | 127 | |
| Logical reasoning | LogiQA | Accuracy38.4 | 84 | |
| General Knowledge Assessment | C-Eval | Accuracy78.5 | 37 | |
| Discrete reasoning | DROP | Exact Match (EM)2.5 | 19 | |
| Lifelong Knowledge Editing | zsRE | Reliability73.5 | 14 | |
| Lifelong Knowledge Editing | CounterFact | Reliability67.1 | 14 | |
| Lifelong Knowledge Editing | QAEdit | Reliability75.1 | 14 | |
| Knowledge Editing | QAEdit | Reliability75.1 | 8 | |
| Conversational Question Answering | CoQA | EM55.1 | 8 | |
| Multi-hop Knowledge Editing | MQuAKE CF v2 | 2-hop Score24.2 | 6 |