Attention-over-Attention Neural Networks for Reading Comprehension
About
Cloze-style queries are representative problems in reading comprehension. Over the past few months, we have seen much progress that utilizing neural network approach to solve Cloze-style questions. In this paper, we present a novel model called attention-over-attention reader for the Cloze-style reading comprehension task. Our model aims to place another attention mechanism over the document-level attention, and induces "attended attention" for final predictions. Unlike the previous works, our neural network model requires less pre-defined hyper-parameters and uses an elegant architecture for modeling. Experimental results show that the proposed attention-over-attention model significantly outperforms various state-of-the-art systems by a large margin in public datasets, such as CNN and Children's Book Test datasets.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Question Answering | SQuAD v1.1 (test) | F1 Score83.8 | 260 | |
| Machine Comprehension | CNN (val) | Accuracy0.731 | 80 | |
| Machine Comprehension | CNN (test) | Accuracy74.4 | 77 | |
| Machine Comprehension | CBT NE (test) | Accuracy72 | 56 | |
| Machine Comprehension | CBT-CN (test) | Accuracy69.4 | 56 | |
| Machine Comprehension | CBT-NE (val) | Accuracy77.8 | 37 | |
| Machine Comprehension | CBT-CN (val) | Accuracy72.2 | 37 | |
| Question Answering | CNN (test) | Accuracy74.4 | 24 | |
| Machine Comprehension | CBT (test) | Named Entities72 | 12 | |
| Reading Comprehension | Children's Book Test (CBT) Named Entity (NE) (dev) | Accuracy77.8 | 12 |