Espresso: A Fast End-to-end Neural Speech Recognition Toolkit
About
We present Espresso, an open-source, modular, extensible end-to-end neural automatic speech recognition (ASR) toolkit based on the deep learning library PyTorch and the popular neural machine translation toolkit fairseq. Espresso supports distributed training across GPUs and computing nodes, and features various decoding approaches commonly employed in ASR, including look-ahead word-based language model fusion, for which a fast, parallelized decoder is implemented. Espresso achieves state-of-the-art ASR performance on the WSJ, LibriSpeech, and Switchboard data sets among other end-to-end systems without data augmentation, and is 4--11x faster for decoding than similar systems (e.g. ESPnet).
Yiming Wang, Tongfei Chen, Hainan Xu, Shuoyang Ding, Hang Lv, Yiwen Shao, Nanyun Peng, Lei Xie, Shinji Watanabe, Sanjeev Khudanpur• 2019
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Automatic Speech Recognition | LibriSpeech (test-other) | WER8.7 | 966 | |
| Automatic Speech Recognition | LibriSpeech clean (test) | WER2.8 | 833 | |
| Automatic Speech Recognition | LibriSpeech (dev-other) | WER8.4 | 411 | |
| Automatic Speech Recognition | LibriSpeech (dev-clean) | WER (%)2.8 | 319 | |
| Speech Recognition | WSJ (92-eval) | WER3.4 | 131 | |
| Speech Recognition | Wall Street Journal open vocabulary (dev93) | WER5.9 | 28 | |
| Automatic Speech Recognition | 80-hour WSJ (dev93) | WER5.9 | 16 | |
| Speech Recognition | Hub5'00 Full (test) | -- | 6 |
Showing 8 of 8 rows