Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Input-to-Output Gate to Improve RNN Language Models

About

This paper proposes a reinforcing method that refines the output layers of existing Recurrent Neural Network (RNN) language models. We refer to our proposed method as Input-to-Output Gate (IOG). IOG has an extremely simple structure, and thus, can be easily combined with any RNN language models. Our experiments on the Penn Treebank and WikiText-2 datasets demonstrate that IOG consistently boosts the performance of several different types of current topline RNN language models.

Sho Takase, Jun Suzuki, Masaaki Nagata• 2017

Related benchmarks

TaskDatasetResultRank
Language ModelingWikiText-2 (test)
PPL91
1541
Language ModelingPenn Treebank (test)
Perplexity64.4
411
Language ModelingWikiText2 (val)
Perplexity (PPL)95.9
277
Language ModelingPTB (val)
Perplexity67
83
Showing 4 of 4 rows

Other info

Follow for update