Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Jasper: An End-to-End Convolutional Neural Acoustic Model

About

In this paper, we report state-of-the-art results on LibriSpeech among end-to-end speech recognition models without any external training data. Our model, Jasper, uses only 1D convolutions, batch normalization, ReLU, dropout, and residual connections. To improve training, we further introduce a new layer-wise optimizer called NovoGrad. Through experiments, we demonstrate that the proposed deep architecture performs as well or better than more complex choices. Our deepest Jasper variant uses 54 convolutional layers. With this architecture, we achieve 2.95% WER using a beam-search decoder with an external neural language model and 3.86% WER with a greedy decoder on LibriSpeech test-clean. We also report competitive results on the Wall Street Journal and the Hub5'00 conversational evaluation datasets.

Jason Li, Vitaly Lavrukhin, Boris Ginsburg, Ryan Leary, Oleksii Kuchaiev, Jonathan M. Cohen, Huyen Nguyen, Ravi Teja Gadde• 2019

Related benchmarks

TaskDatasetResultRank
Automatic Speech RecognitionLibriSpeech (test-other)
WER7.84
966
Automatic Speech RecognitionLibriSpeech clean (test)
WER2.84
833
Automatic Speech RecognitionLibriSpeech (dev-other)
WER7.61
411
Automatic Speech RecognitionLibriSpeech (dev-clean)
WER (%)2.62
319
Speech RecognitionWSJ (92-eval)
WER6.9
131
Speech RecognitionHub5'00
SWB Score7.8
25
Speech RecognitionWSJ 93 (test)
WER9.3
13
Showing 7 of 7 rows

Other info

Code

Follow for update