Clues Before Answers: Generation-Enhanced Multiple-Choice QA
About
A trending paradigm for multiple-choice question answering (MCQA) is using a text-to-text framework. By unifying data in different tasks into a single text-to-text format, it trains a generative encoder-decoder model which is both powerful and universal. However, a side effect of twisting a generation target to fit the classification nature of MCQA is the under-utilization of the decoder and the knowledge that can be decoded. To exploit the generation capability and underlying knowledge of a pre-trained encoder-decoder model, in this paper, we propose a generation-enhanced MCQA model named GenMC. It generates a clue from the question and then leverages the clue to enhance a reader for MCQA. It outperforms text-to-text models on multiple MCQA datasets.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Question Answering | OpenBookQA (OBQA) (test) | OBQA Accuracy66.87 | 130 | |
| Commonsense Question Answering | CSQA (test) | Accuracy0.7267 | 127 | |
| Multiple-choice Question Answering | ARC Easy (test) | Accuracy0.6901 | 50 | |
| Multiple-choice Question Answering | ARC Challenge (test) | Accuracy47.41 | 26 | |
| Multiple-choice Question Answering | OBQA (dev) | Accuracy71.6 | 17 | |
| Multiple-choice Question Answering | QASC (test) | Accuracy58.06 | 16 | |
| Question Answering | OpenBookQA Official Leaderboard | Accuracy92 | 14 | |
| Multiple-choice Question Answering | CSQA (dev) | Accuracy71.1 | 10 | |
| Multiple-choice Question Answering | ARC Easy (dev) | Accuracy72.49 | 10 | |
| Multiple-choice Question Answering | QASC (dev) | Accuracy67.61 | 10 |