Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Self-Knowledge Guided Retrieval Augmentation for Large Language Models

About

Large language models (LLMs) have shown superior performance without task-specific fine-tuning. Despite the success, the knowledge stored in the parameters of LLMs could still be incomplete and difficult to update due to the computational costs. As complementary, retrieval-based methods can offer non-parametric world knowledge and improve the performance on tasks such as question answering. However, we find that the retrieved knowledge does not always help and even has a negative impact on original responses occasionally. To better make use of both internal knowledge and external world knowledge, we investigate eliciting the model's ability to recognize what they know and do not know (which is also called self-knowledge) and propose Self-Knowledge guided Retrieval augmentation (SKR), a simple yet effective method which can let LLMs refer to the questions they have previously encountered and adaptively call for external resources when dealing with new questions. We evaluate SKR on multiple datasets and demonstrate that it outperforms chain-of-thought based and fully retrieval-based methods by using either InstructGPT or ChatGPT.

Yile Wang, Peng Li, Maosong Sun, Yang Liu• 2023

Related benchmarks

TaskDatasetResultRank
Multi-hop Question Answering2WikiMultihopQA
EM32.5
278
Multi-hop Question AnsweringHotpotQA--
221
Open-domain Question AnsweringTriviaQA
EM67.5
62
Question AnsweringStrategyQA
EM70.1
35
Open-domain Question AnsweringNQ (Natural Questions)
EM33.8
33
Question AnsweringNaturalQA
EM28.53
26
Question AnsweringWebQA
EM24.16
23
Question AnsweringAverage (NQ, TriviaQA, HotpotQA, StrategyQA, 2WikiMHQA)
Average Score47.6
14
Multi-hop Question AnsweringHotpotQA in-domain (test)
ACC_R38
14
Multi-hop Question Answering2WikiMultiHopQA in-domain (test)
Accuracy (Response)27.8
14
Showing 10 of 15 rows

Other info

Follow for update