Densely Connected Attention Propagation for Reading Comprehension
About
We propose DecaProp (Densely Connected Attention Propagation), a new densely connected neural architecture for reading comprehension (RC). There are two distinct characteristics of our model. Firstly, our model densely connects all pairwise layers of the network, modeling relationships between passage and query across all hierarchical levels. Secondly, the dense connectors in our network are learned via attention instead of standard residual skip-connectors. To this end, we propose novel Bidirectional Attention Connectors (BAC) for efficiently forging connections throughout the network. We conduct extensive experiments on four challenging RC benchmarks. Our proposed approach achieves state-of-the-art results on all four, outperforming existing baselines by up to $2.6\%-14.2\%$ in absolute F1 score.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Question Answering | NewsQA (dev) | F1 Score65.7 | 101 | |
| Question Answering | SearchQA (test) | N-gram F170.8 | 48 | |
| Open-domain Question Answering | Quasar-T (test) | F1 Score46.9 | 33 | |
| Question Answering | NewsQA (test) | F166.3 | 31 | |
| Question Answering | SearchQA (dev) | -- | 28 | |
| Reading Comprehension | SQuAD (dev) | F1 Score0.814 | 15 | |
| Reading Comprehension | NarrativeQA (test) | BLEU-142 | 8 | |
| Reading Comprehension | Quasar-T (test) | EM38.6 | 6 | |
| Reading Comprehension | NarrativeQA Story Summaries (val) | BLEU-144.35 | 6 | |
| Reading Comprehension | NarrativeQA Story Summaries (test) | BLEU-142 | 4 |