Recurrent Dropout without Memory Loss
About
This paper presents a novel approach to recurrent neural network (RNN) regularization. Differently from the widely adopted dropout method, which is applied to \textit{forward} connections of feed-forward architectures or RNNs, we propose to drop neurons directly in \textit{recurrent} connections in a way that does not cause loss of long-term memory. Our approach is as easy to implement and apply as the regular feed-forward dropout and we demonstrate its effectiveness for Long Short-Term Memory network, the most popular type of RNN cells. Our experiments on NLP benchmarks show consistent improvements even when combined with conventional feed-forward dropout.
Stanislau Semeniuta, Aliaksei Severyn, Erhardt Barth• 2016
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Character-level Language Modeling | Penn Treebank (test) | BPC1.3 | 113 | |
| Sequential Image Classification | PMNIST (test) | Accuracy (Test)92.5 | 77 | |
| Language Modeling | Penn Treebank word-level (test) | Perplexity87 | 72 | |
| Character-level Language Modeling | Penn Treebank char-level (test) | BPC1.32 | 25 | |
| Character-level Language Modeling | Penn Treebank character-level (val) | BPC1.3 | 10 |
Showing 5 of 5 rows