Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

End-To-End Memory Networks

About

We introduce a neural network with a recurrent attention model over a possibly large external memory. The architecture is a form of Memory Network (Weston et al., 2015) but unlike the model in that work, it is trained end-to-end, and hence requires significantly less supervision during training, making it more generally applicable in realistic settings. It can also be seen as an extension of RNNsearch to the case where multiple computational steps (hops) are performed per output symbol. The flexibility of the model allows us to apply it to tasks as diverse as (synthetic) question answering and to language modeling. For the former our approach is competitive with Memory Networks, but with less supervision. For the latter, on the Penn TreeBank and Text8 datasets our approach demonstrates comparable performance to RNNs and LSTMs. In both cases we show that the key concept of multiple computational hops yields improved results.

Sainbayar Sukhbaatar, Arthur Szlam, Jason Weston, Rob Fergus• 2015

Related benchmarks

TaskDatasetResultRank
Language ModelingPenn Treebank (test)
Perplexity111
411
Language ModelingPenn Treebank (val)
Perplexity118
178
Emotion Recognition in ConversationIEMOCAP (test)--
154
Answer SelectionWikiQA (test)
MAP0.517
149
Conversational Emotion RecognitionIEMOCAP
Weighted Average F1 Score55.1
129
Question AnsweringbAbI (test)
Mean Error0.042
54
Multimodal Emotion Recognition in ConversationIEMOCAP 6-class (test)
Weighted F1 Score (WF1)55.1
33
Multimodal Emotion RecognitionIEMOCAP 6-way
F1 (Avg)59.5
28
DialogbAbI dialog 1.0 (OOV)
Avg Error Rate11.2
22
Language ModelingText8 word-level (test)
Perplexity147
19
Showing 10 of 35 rows

Other info

Code

Follow for update