From Cloze to Comprehension: Retrofitting Pre-trained Masked Language Model to Pre-trained Machine Reader
About
We present Pre-trained Machine Reader (PMR), a novel method for retrofitting pre-trained masked language models (MLMs) to pre-trained machine reading comprehension (MRC) models without acquiring labeled data. PMR can resolve the discrepancy between model pre-training and downstream fine-tuning of existing MLMs. To build the proposed PMR, we constructed a large volume of general-purpose and high-quality MRC-style training data by using Wikipedia hyperlinks and designed a Wiki Anchor Extraction task to guide the MRC-style pre-training. Apart from its simplicity, PMR effectively solves extraction tasks, such as Extractive Question Answering and Named Entity Recognition. PMR shows tremendous improvements over existing approaches, especially in low-resource scenarios. When applied to the sequence classification task in the MRC formulation, PMR enables the extraction of high-quality rationales to explain the classification process, thereby providing greater prediction explainability. PMR also has the potential to serve as a unified model for tackling various extraction and classification tasks in the MRC formulation.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Question Answering | NewsQA (dev) | F1 Score52.3 | 101 | |
| Question Answering | SQuAD (dev) | F179.8 | 74 | |
| Question Answering | Natural Question (NQ) (dev) | F157.4 | 72 | |
| Question Answering | HotpotQA (dev) | Answer F165.9 | 43 | |
| Question Answering | TRIVIAQA (dev) | F1 Score (Full)68.6 | 32 | |
| Question Answering | SearchQA (dev) | F1 (N-gram)68.5 | 28 | |
| Question Answering | BioASQ (dev) | F1 Score76.8 | 28 | |
| Question Answering | TextbookQA (dev) | F1 Score45.1 | 28 | |
| Named Entity Recognition | CoNLL (test) | -- | 28 |