Pushing the Limits of Semi-Supervised Learning for Automatic Speech Recognition
About
We employ a combination of recent developments in semi-supervised learning for automatic speech recognition to obtain state-of-the-art results on LibriSpeech utilizing the unlabeled audio of the Libri-Light dataset. More precisely, we carry out noisy student training with SpecAugment using giant Conformer models pre-trained using wav2vec 2.0 pre-training. By doing so, we are able to achieve word-error-rates (WERs) 1.4%/2.6% on the LibriSpeech test/test-other sets against the current state-of-the-art WERs 1.7%/3.3%.
Yu Zhang, James Qin, Daniel S. Park, Wei Han, Chung-Cheng Chiu, Ruoming Pang, Quoc V. Le, Yonghui Wu• 2020
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Automatic Speech Recognition | LibriSpeech (test-other) | WER2.6 | 966 | |
| Automatic Speech Recognition | LibriSpeech clean (test) | WER1.4 | 833 | |
| Automatic Speech Recognition | LibriSpeech (dev-other) | WER2.6 | 411 | |
| Automatic Speech Recognition | LibriSpeech (dev-clean) | WER (%)1.3 | 319 | |
| Automatic Speech Recognition | LibriSpeech 960h (test-other) | WER2.6 | 81 | |
| Speech Recognition | LibriSpeech (test) | WER0.014 | 59 | |
| Speech Recognition | LibriSpeech clean (dev) | WER0.02 | 59 | |
| Automatic Speech Recognition | LibriSpeech 960h (dev-other) | WER2.6 | 50 | |
| Speech Recognition | LibriSpeech 960hr (test) | WER1.4 | 26 | |
| Speech Recognition | LibriSpeech 960hr (dev) | WER1.3 | 25 |
Showing 10 of 13 rows