Memory Networks
About
We describe a new class of learning models called memory networks. Memory networks reason with inference components combined with a long-term memory component; they learn how to use these jointly. The long-term memory can be read and written to, with the goal of using it for prediction. We investigate these models in the context of question answering (QA) where the long-term memory effectively acts as a (dynamic) knowledge base, and the output is a textual response. We evaluate them on a large-scale QA task, and a smaller, but more complex, toy task generated from a simulated world. In the latter, we show the reasoning power of such models by chaining multiple supporting sentences to answer questions that require understanding the intension of verbs.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Machine Comprehension | CNN (val) | Accuracy0.662 | 80 | |
| Machine Comprehension | CNN (test) | Accuracy69.4 | 77 | |
| Machine Comprehension | CBT NE (test) | Accuracy66.6 | 56 | |
| Machine Comprehension | CBT-CN (test) | Accuracy63 | 56 | |
| Question Answering | bAbI (test) | Mean Error2.81 | 54 | |
| Machine Comprehension | CBT-NE (val) | Accuracy70.4 | 37 | |
| Machine Comprehension | CBT-CN (val) | Accuracy64.2 | 37 | |
| Question Answering | Reverb (test) | Accuracy72 | 15 | |
| Question Answering | bAbI 10k (test) | Task 1: 1 Supporting Fact Error0.00e+0 | 15 | |
| Machine Comprehension | CBT (test) | Named Entities49.3 | 12 |