Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

On Multiplicative Integration with Recurrent Neural Networks

About

We introduce a general and simple structural design called Multiplicative Integration (MI) to improve recurrent neural networks (RNNs). MI changes the way in which information from difference sources flows and is integrated in the computational building block of an RNN, while introducing almost no extra parameters. The new structure can be easily embedded into many popular RNN models, including LSTMs and GRUs. We empirically analyze its learning behaviour and conduct evaluations on several tasks using different RNN models. Our experimental results demonstrate that Multiplicative Integration can provide a substantial performance boost over many of the existing RNN models.

Yuhuai Wu, Saizheng Zhang, Ying Zhang, Yoshua Bengio, Ruslan Salakhutdinov• 2016

Related benchmarks

TaskDatasetResultRank
Character-level Language Modelingenwik8 (test)
BPC1.44
195
Speech RecognitionWSJ (92-eval)
WER8.2
131
Character-level Language Modelingtext8 (test)
BPC1.44
128
Character-level Language ModelingPenn Treebank (test)
BPC1.39
113
Character-level Language ModelingHutter Prize Wikipedia (test)
Bits/Char1.44
28
Byte-size token predictionByte-size token prediction dataset (val)
BPC1.44
7
Showing 6 of 6 rows

Other info

Follow for update