Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Atlas: Few-shot Learning with Retrieval Augmented Language Models

About

Large language models have shown impressive few-shot results on a wide range of tasks. However, when knowledge is key for such results, as is the case for tasks such as question answering and fact checking, massive parameter counts to store knowledge seem to be needed. Retrieval augmented models are known to excel at knowledge intensive tasks without the need for as many parameters, but it is unclear whether they work in few-shot settings. In this work we present Atlas, a carefully designed and pre-trained retrieval augmented language model able to learn knowledge intensive tasks with very few training examples. We perform evaluations on a wide range of tasks, including MMLU, KILT and NaturalQuestions, and study the impact of the content of the document index, showing that it can easily be updated. Notably, Atlas reaches over 42% accuracy on Natural Questions using only 64 examples, outperforming a 540B parameters model by 3% despite having 50x fewer parameters.

Gautier Izacard, Patrick Lewis, Maria Lomeli, Lucas Hosseini, Fabio Petroni, Timo Schick, Jane Dwivedi-Yu, Armand Joulin, Sebastian Riedel, Edouard Grave• 2022

Related benchmarks

TaskDatasetResultRank
Multitask Language UnderstandingMMLU (test)
Accuracy66
303
Multi-hop Question Answering2WikiMultihopQA
EM38.1
278
Multi-hop Question AnsweringHotpotQA
F1 Score55.3
221
Question AnsweringTriviaQA--
85
Fact VerificationFEVER
Accuracy0.77
67
Multi-hop Question AnsweringMulti-hop RAG
F152.3
65
Question AnsweringNQ (Natural Questions)
EM26.7
55
Open-domain Question AnsweringNaturalQ-Open (test)
EM60.4
37
Question AnsweringNatural Questions (NQ) (test)
Exact Match60.4
35
Fact VerificationFEVER (test)--
32
Showing 10 of 26 rows

Other info

Code

Follow for update