Words or Characters? Fine-grained Gating for Reading Comprehension
About
Previous work combines word-level and character-level representations using concatenation or scalar weighting, which is suboptimal for high-level tasks like reading comprehension. We present a fine-grained gating mechanism to dynamically combine word-level and character-level representations based on properties of the words. We also extend the idea of fine-grained gating to modeling the interaction between questions and paragraphs for reading comprehension. Experiments show that our approach can improve the performance on reading comprehension tasks, achieving new state-of-the-art results on the Children's Book Test dataset. To demonstrate the generality of our gating mechanism, we also show improved results on a social media tag prediction task.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Question Answering | SQuAD v1.1 (test) | F1 Score73.327 | 260 | |
| Question Answering | SQuAD (test) | F173.3 | 111 | |
| Question Answering | SQuAD (dev) | F173.41 | 74 | |
| Machine Comprehension | CBT NE (test) | Accuracy74.96 | 56 | |
| Machine Comprehension | CBT-CN (test) | Accuracy72.04 | 56 | |
| Reading Comprehension | Children's Book Test (CBT) Common Noun (CN) (dev) | Accuracy75.3 | 12 | |
| Reading Comprehension | Children's Book Test (CBT) Named Entity (NE) (dev) | Accuracy79.1 | 12 | |
| Tag prediction | Twitter dataset | Precision@130.69 | 6 | |
| Reading Comprehension | Who Did What (WDW) (test) | Accuracy71.7 | 5 | |
| Reading Comprehension | Who Did What Relaxed (WDW-R) (test) | Accuracy72.6 | 5 |