Robust Speech Recognition via Large-Scale Weak Supervision
About
We study the capabilities of speech processing systems trained simply to predict large amounts of transcripts of audio on the internet. When scaled to 680,000 hours of multilingual and multitask supervision, the resulting models generalize well to standard benchmarks and are often competitive with prior fully supervised results but in a zero-shot transfer setting without the need for any fine-tuning. When compared to humans, the models approach their accuracy and robustness. We are releasing models and inference code to serve as a foundation for further work on robust speech processing.
Alec Radford, Jong Wook Kim, Tao Xu, Greg Brockman, Christine McLeavey, Ilya Sutskever• 2022
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Automatic Speech Recognition | LibriSpeech (test-other) | WER3.5 | 966 | |
| Automatic Speech Recognition | LibriSpeech clean (test) | WER1.82 | 833 | |
| Automatic Speech Recognition | Librispeech (test-clean) | WER3.21 | 84 | |
| Automatic Speech Recognition | LibriSpeech 960h (test-other) | WER6.01 | 81 | |
| Automatic Speech Recognition | LibriSpeech Other | WER3.55 | 75 | |
| Automatic Speech Recognition | AISHELL-1 (test) | CER514 | 71 | |
| Audio-Visual Speech Recognition | LRS3 clean (test) | WER2.1 | 70 | |
| Speech Recognition | LibriSpeech (test) | WER0.027 | 59 | |
| Multimodal Sentiment Analysis | CMU-MOSI | -- | 59 | |
| Automatic Speech Recognition | LibriSpeech Clean | WER1.51 | 57 |
Showing 10 of 307 rows
...