Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Dynamic Evaluation of Neural Sequence Models

About

We present methodology for using dynamic evaluation to improve neural sequence models. Models are adapted to recent history via a gradient descent based mechanism, causing them to assign higher probabilities to re-occurring sequential patterns. Dynamic evaluation outperforms existing adaptation approaches in our comparisons. Dynamic evaluation improves the state-of-the-art word-level perplexities on the Penn Treebank and WikiText-2 datasets to 51.1 and 44.3 respectively, and the state-of-the-art character-level cross-entropies on the text8 and Hutter Prize datasets to 1.19 bits/char and 1.08 bits/char respectively.

Ben Krause, Emmanuel Kahembwe, Iain Murray, Steve Renals• 2017

Related benchmarks

TaskDatasetResultRank
Language ModelingWikiText-2 (test)
PPL44.3
1541
Language ModelingPenn Treebank (test)
Perplexity51.1
411
Language ModelingWikiText2 v1 (test)
Perplexity44.3
341
Language ModelingWikiText2 (val)
Perplexity (PPL)46.4
277
Character-level Language Modelingenwik8 (test)
BPC1.08
195
Language ModelingPenn Treebank (val)
Perplexity51.6
178
Character-level Language Modelingtext8 (test)
BPC1.19
128
Character-level Language ModelingHutter Prize Wikipedia (test)
Bits/Char1.08
28
Language ModelingWikiText-2 v1 (val)
Perplexity46.4
19
Showing 9 of 9 rows

Other info

Code

Follow for update