Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Adaptive Input Representations for Neural Language Modeling

About

We introduce adaptive input representations for neural language modeling which extend the adaptive softmax of Grave et al. (2017) to input representations of variable capacity. There are several choices on how to factorize the input and output layers, and whether to model words, characters or sub-word units. We perform a systematic comparison of popular choices for a self-attentional architecture. Our experiments show that models equipped with adaptive embeddings are more than twice as fast to train than the popular character input CNN while having a lower number of parameters. On the WikiText-103 benchmark we achieve 18.7 perplexity, an improvement of 10.5 perplexity compared to the previously best published result and on the Billion Word benchmark, we achieve 23.02 perplexity.

Alexei Baevski, Michael Auli• 2018

Related benchmarks

TaskDatasetResultRank
Automatic Speech RecognitionLibriSpeech clean (test)
WER4.18
1156
Automatic Speech RecognitionLibriSpeech (test-other)
WER9.18
1151
Language ModelingWikiText-103 (test)
Perplexity18.5
579
Natural Language UnderstandingGLUE
SST-292.31
531
Language ModelingPTB (test)
Perplexity57
526
Automatic Speech RecognitionLibriSpeech (dev-other)
WER9.06
462
Language ModelingWikiText-103 (val)
PPL17.97
214
Long-range sequence modelingLong Range Arena (LRA)
Text Accuracy65.21
177
Machine TranslationIWSLT De-En 2014 (test)
BLEU35.9
146
Language ModelingOne Billion Word Benchmark (test)
Test Perplexity23.02
113
Showing 10 of 65 rows

Other info

Code

Follow for update