Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

In-Context Retrieval-Augmented Language Models

About

Retrieval-Augmented Language Modeling (RALM) methods, which condition a language model (LM) on relevant documents from a grounding corpus during generation, were shown to significantly improve language modeling performance. In addition, they can mitigate the problem of factually inaccurate text generation and provide natural source attribution mechanism. Existing RALM approaches focus on modifying the LM architecture in order to facilitate the incorporation of external information, significantly complicating deployment. This paper considers a simple alternative, which we dub In-Context RALM: leaving the LM architecture unchanged and prepending grounding documents to the input, without any further training of the LM. We show that In-Context RALM that builds on off-the-shelf general purpose retrievers provides surprisingly large LM gains across model sizes and diverse corpora. We also demonstrate that the document retrieval and ranking mechanism can be specialized to the RALM setting to further boost performance. We conclude that In-Context RALM has considerable potential to increase the prevalence of LM grounding, particularly in settings where a pretrained LM must be used without modification or even via API access.

Ori Ram, Yoav Levine, Itay Dalmedigos, Dor Muhlgay, Amnon Shashua, Kevin Leyton-Brown, Yoav Shoham• 2023

Related benchmarks

TaskDatasetResultRank
Multi-hop Question Answering2WikiMultihopQA
EM30.8
278
Multi-hop Question AnsweringHotpotQA
F1 Score50.88
221
Question AnsweringTriviaQA
Accuracy82.85
210
Question AnsweringPopQA
Accuracy36.97
186
Multi-hop Question Answering2WikiMQA
F1 Score44.69
154
Question AnsweringTriviaQA (test)
Accuracy82.85
121
Question AnsweringNQ
Accuracy47.66
108
Multi-hop Question AnsweringHotpotQA
F154.73
79
Question AnsweringNQ (test)--
66
Multi-hop Question AnsweringMulti-hop RAG--
65
Showing 10 of 39 rows

Other info

Follow for update