Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Distilling Knowledge from Reader to Retriever for Question Answering

About

The task of information retrieval is an important component of many natural language processing systems, such as open domain question answering. While traditional methods were based on hand-crafted features, continuous representations based on neural networks recently obtained competitive results. A challenge of using such methods is to obtain supervised data to train the retriever model, corresponding to pairs of query and support documents. In this paper, we propose a technique to learn retriever models for downstream tasks, inspired by knowledge distillation, and which does not require annotated pairs of query and documents. Our approach leverages attention scores of a reader model, used to solve the task based on retrieved documents, to obtain synthetic labels for the retriever. We evaluate our method on question answering, obtaining state-of-the-art results.

Gautier Izacard, Edouard Grave• 2020

Related benchmarks

TaskDatasetResultRank
Open Question AnsweringNatural Questions (NQ) (test)
Exact Match (EM)53.7
134
Open-domain Question AnsweringTriviaQA (test)
Exact Match72.5
80
Passage retrievalTriviaQA (test)
Top-100 Acc87.7
67
Question AnsweringNQ (test)
EM Accuracy54.4
66
RetrievalNatural Questions (test)
Top-5 Recall73.8
62
Question AnsweringNarrativeQA (test)
ROUGE-L32
61
Open-domain Question AnsweringWebQuestions (WebQ) (test)
Exact Match (EM)46.8
55
End-to-end Open-Domain Question AnsweringNQ (test)
Exact Match (EM)50.06
50
Open-domain Question AnsweringNaturalQ-Open (test)
EM54.7
37
Open-domain Question AnsweringTriviaQA (TQA) (test)
Accuracy72.1
26
Showing 10 of 26 rows

Other info

Code

Follow for update