Iterative Alternating Neural Attention for Machine Reading
About
We propose a novel neural attention architecture to tackle machine comprehension tasks, such as answering Cloze-style queries with respect to a document. Unlike previous models, we do not collapse the query into a single vector, instead we deploy an iterative alternating attention mechanism that allows a fine-grained exploration of both the query and the document. Our model outperforms state-of-the-art baselines in standard machine comprehension benchmarks such as CNN news articles and the Children's Book Test (CBT) dataset.
Alessandro Sordoni, Philip Bachman, Adam Trischler, Yoshua Bengio• 2016
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Machine Comprehension | CNN (val) | Accuracy0.752 | 80 | |
| Machine Comprehension | CNN (test) | Accuracy76.1 | 77 | |
| Machine Comprehension | CBT-CN (test) | Accuracy71 | 56 | |
| Machine Comprehension | CBT NE (test) | Accuracy72 | 56 | |
| Machine Comprehension | CBT-CN (val) | Accuracy74.1 | 37 | |
| Machine Comprehension | CBT-NE (val) | Accuracy76.9 | 37 | |
| Question Answering | CNN (test) | Accuracy75.7 | 24 | |
| Reading Comprehension | Children's Book Test (CBT) Common Noun (CN) (dev) | Accuracy74.1 | 12 | |
| Reading Comprehension | Children's Book Test (CBT) Named Entity (NE) (dev) | Accuracy76.9 | 12 |
Showing 9 of 9 rows