Self-Attentive Associative Memory
About
Heretofore, neural networks with external memory are restricted to single memory with lossy representations of memory interactions. A rich representation of relationships between memory pieces urges a high-order and segregated relational memory. In this paper, we propose to separate the storage of individual experiences (item memory) and their occurring relationships (relational memory). The idea is implemented through a novel Self-attentive Associative Memory (SAM) operator. Found upon outer product, SAM forms a set of associative memories that represent the hypothetical high-order relationships between arbitrary pairs of memory elements, through which a relational memory is constructed from an item memory. The two memories are wired into a single sequential model capable of both memorization and relational reasoning. We achieve competitive results with our proposed two-memory model in a diversity of machine learning tasks, from challenging synthetic problems to practical testbeds such as geometry, graph, reinforcement learning, and question answering.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Question Answering | bAbI (test) | Mean Error0.39 | 54 | |
| sys-bAbI task | sys-bAbI original (test) | Gap3.7 | 22 | |
| Nth-farthest | Nth-farthest (test) | Accuracy98 | 6 | |
| Multi-hop spatial reasoning | StepGame with distracting noise (test) | k=1 Accuracy53.42 | 6 | |
| Multi-hop spatial reasoning | StepGame larger k generalization (test) | Accuracy (k=6)13.8 | 6 | |
| Spatial Reasoning | bAbI original (test) | Task 17 Accuracy97.8 | 6 |