Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

FAME: Towards Factual Multi-Task Model Editing

About

Large language models (LLMs) embed extensive knowledge and utilize it to perform exceptionally well across various tasks. Nevertheless, outdated knowledge or factual errors within LLMs can lead to misleading or incorrect responses, causing significant issues in practical applications. To rectify the fatal flaw without the necessity for costly model retraining, various model editing approaches have been proposed to correct inaccurate knowledge within LLMs in a cost-efficient way. To evaluate these model editing methods, previous work introduced a series of datasets. However, most of the previous datasets only contain fabricated data in a single format, which diverges from real-world model editing scenarios, raising doubts about their usability in practice. To facilitate the application of model editing in real-world scenarios, we propose the challenge of practicality. To resolve such challenges and effectively enhance the capabilities of LLMs, we present FAME, an factual, comprehensive, and multi-task dataset, which is designed to enhance the practicality of model editing. We then propose SKEME, a model editing method that uses a novel caching mechanism to ensure synchronization with the real world. The experiments demonstrate that SKEME performs excellently across various tasks and scenarios, confirming its practicality.

Li Zeng, Yingyu Shan, Zeming Liu, Jiashu Yao, Yuhang Guo• 2024

Related benchmarks

TaskDatasetResultRank
Knowledge EditingCounterFact
Efficacy85.83
301
Knowledge EditingzsRE--
181
Knowledge EditingCounterfact uns
Edit Success Rate82.85
30
Knowledge EditingWikiUpdate
Edit Success71.84
30
Knowledge EditingMQuAKE
Edit Success Rate90.98
30
Showing 5 of 5 rows

Other info

Follow for update