Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Attention-over-Attention Neural Networks for Reading Comprehension

About

Cloze-style queries are representative problems in reading comprehension. Over the past few months, we have seen much progress that utilizing neural network approach to solve Cloze-style questions. In this paper, we present a novel model called attention-over-attention reader for the Cloze-style reading comprehension task. Our model aims to place another attention mechanism over the document-level attention, and induces "attended attention" for final predictions. Unlike the previous works, our neural network model requires less pre-defined hyper-parameters and uses an elegant architecture for modeling. Experimental results show that the proposed attention-over-attention model significantly outperforms various state-of-the-art systems by a large margin in public datasets, such as CNN and Children's Book Test datasets.

Yiming Cui, Zhipeng Chen, Si Wei, Shijin Wang, Ting Liu, Guoping Hu• 2016

Related benchmarks

TaskDatasetResultRank
Question AnsweringSQuAD v1.1 (test)
F1 Score83.8
260
Machine ComprehensionCNN (val)
Accuracy0.731
80
Machine ComprehensionCNN (test)
Accuracy74.4
77
Machine ComprehensionCBT NE (test)
Accuracy72
56
Machine ComprehensionCBT-CN (test)
Accuracy69.4
56
Machine ComprehensionCBT-NE (val)
Accuracy77.8
37
Machine ComprehensionCBT-CN (val)
Accuracy72.2
37
Question AnsweringCNN (test)
Accuracy74.4
24
Machine ComprehensionCBT (test)
Named Entities72
12
Reading ComprehensionChildren's Book Test (CBT) Named Entity (NE) (dev)
Accuracy77.8
12
Showing 10 of 14 rows

Other info

Follow for update