Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Recurrent Dropout without Memory Loss

About

This paper presents a novel approach to recurrent neural network (RNN) regularization. Differently from the widely adopted dropout method, which is applied to \textit{forward} connections of feed-forward architectures or RNNs, we propose to drop neurons directly in \textit{recurrent} connections in a way that does not cause loss of long-term memory. Our approach is as easy to implement and apply as the regular feed-forward dropout and we demonstrate its effectiveness for Long Short-Term Memory network, the most popular type of RNN cells. Our experiments on NLP benchmarks show consistent improvements even when combined with conventional feed-forward dropout.

Stanislau Semeniuta, Aliaksei Severyn, Erhardt Barth• 2016

Related benchmarks

TaskDatasetResultRank
Character-level Language ModelingPenn Treebank (test)
BPC1.3
113
Sequential Image ClassificationPMNIST (test)
Accuracy (Test)92.5
77
Language ModelingPenn Treebank word-level (test)
Perplexity87
72
Character-level Language ModelingPenn Treebank char-level (test)
BPC1.32
25
Character-level Language ModelingPenn Treebank character-level (val)
BPC1.3
10
Showing 5 of 5 rows

Other info

Follow for update