Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Few-Shot Question Answering by Pretraining Span Selection

About

In several question answering benchmarks, pretrained models have reached human parity through fine-tuning on an order of 100,000 annotated questions and answers. We explore the more realistic few-shot setting, where only a few hundred training examples are available, and observe that standard models perform poorly, highlighting the discrepancy between current pretraining objectives and question answering. We propose a new pretraining scheme tailored for question answering: recurring span selection. Given a passage with multiple sets of recurring spans, we mask in each set all recurring spans but one, and ask the model to select the correct span in the passage for each masked span. Masked spans are replaced with a special token, viewed as a question representation, that is later used during fine-tuning to select the answer span. The resulting model obtains surprisingly good results on multiple benchmarks (e.g., 72.7 F1 on SQuAD with only 128 training examples), while maintaining competitive performance in the high-resource setting.

Ori Ram, Yuval Kirstain, Jonathan Berant, Amir Globerson, Omer Levy• 2021

Related benchmarks

TaskDatasetResultRank
Question AnsweringNewsQA (dev)
F1 Score43.5
101
Question AnsweringSQuAD (dev)
F172.7
74
Question AnsweringNatural Question (NQ) (dev)
F146.3
72
Question AnsweringHotpotQA (dev)
Answer F154.7
43
Question AnsweringMRQA 2019 (dev)
SQuAD Score92.4
32
Question AnsweringTRIVIAQA (dev)
F1 Score (Full)44.7
32
Question AnsweringTextbookQA (dev)
F1 Score42.6
28
Question AnsweringBioASQ (dev)
F1 Score63.2
28
Question AnsweringSearchQA (dev)
F1 (N-gram)47.2
28
Extractive Question AnsweringSearchQA MRQA
F1 Score83
22
Showing 10 of 19 rows

Other info

Code

Follow for update