Simple Recurrent Units for Highly Parallelizable Recurrence
About
Common recurrent neural architectures scale poorly due to the intrinsic difficulty in parallelizing their state computations. In this work, we propose the Simple Recurrent Unit (SRU), a light recurrent unit that balances model capacity and scalability. SRU is designed to provide expressive recurrence, enable highly parallelized implementation, and comes with careful initialization to facilitate training of deep models. We demonstrate the effectiveness of SRU on multiple NLP tasks. SRU achieves 5--9x speed-up over cuDNN-optimized LSTM on classification and question answering datasets, and delivers stronger results than LSTM and convolutional models. We also obtain an average of 0.7 BLEU improvement over the Transformer model on translation by incorporating SRU into the architecture.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Machine Translation | WMT En-De 2014 (test) | BLEU28.4 | 379 | |
| Question Answering | SQuAD v1.1 (dev) | F1 Score80.2 | 375 | |
| Character-level Language Modeling | enwik8 (test) | BPC1.19 | 195 | |
| Machine Translation | WMT English-German 2014 (test) | BLEU28.3 | 136 | |
| Subjectivity Classification | Subj (test) | Accuracy93.8 | 125 | |
| Question Classification | TREC (test) | Accuracy94.8 | 124 | |
| Sentiment Classification | Stanford Sentiment Treebank SST-2 (test) | Accuracy89.6 | 99 | |
| Text Classification | MR (test) | Accuracy83.1 | 99 | |
| Language Modeling | Penn Treebank word-level (test) | Perplexity60.3 | 72 | |
| Sentence Classification | CR (test) | Accuracy86.4 | 33 |