Input-to-Output Gate to Improve RNN Language Models
About
This paper proposes a reinforcing method that refines the output layers of existing Recurrent Neural Network (RNN) language models. We refer to our proposed method as Input-to-Output Gate (IOG). IOG has an extremely simple structure, and thus, can be easily combined with any RNN language models. Our experiments on the Penn Treebank and WikiText-2 datasets demonstrate that IOG consistently boosts the performance of several different types of current topline RNN language models.
Sho Takase, Jun Suzuki, Masaaki Nagata• 2017
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Language Modeling | WikiText-2 (test) | PPL91 | 1541 | |
| Language Modeling | Penn Treebank (test) | Perplexity64.4 | 411 | |
| Language Modeling | WikiText2 (val) | Perplexity (PPL)95.9 | 277 | |
| Language Modeling | PTB (val) | Perplexity67 | 83 |
Showing 4 of 4 rows