Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Shuffling Recurrent Neural Networks

About

We propose a novel recurrent neural network model, where the hidden state $h_t$ is obtained by permuting the vector elements of the previous hidden state $h_{t-1}$ and adding the output of a learned function $b(x_t)$ of the input $x_t$ at time $t$. In our model, the prediction is given by a second learned function, which is applied to the hidden state $s(h_t)$. The method is easy to implement, extremely efficient, and does not suffer from vanishing nor exploding gradients. In an extensive set of experiments, the method shows competitive results, in comparison to the leading literature baselines.

Michael Rotman, Lior Wolf• 2020

Related benchmarks

TaskDatasetResultRank
Permuted Image ClassificationpMNIST 784
Accuracy96.43
7
Permuted Image ClassificationpMNIST 3136
Accuracy90.31
7
Sequence CopyingCopy Length 120
Accuracy100
7
Scattered Sequence CopyingScattered Copy Length 5020
Accuracy0.303
6
Sequence CopyingCopy Length 5020
Accuracy70.8
6
Sequence AdditionAdd Length 5000
MSE0.0121
6
Showing 6 of 6 rows

Other info

Follow for update