Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Large-Scale Self- and Semi-Supervised Learning for Speech Translation

About

In this paper, we improve speech translation (ST) through effectively leveraging large quantities of unlabeled speech and text data in different and complementary ways. We explore both pretraining and self-training by using the large Libri-Light speech audio corpus and language modeling with CommonCrawl. Our experiments improve over the previous state of the art by 2.6 BLEU on average on all four considered CoVoST 2 language pairs via a simple recipe of combining wav2vec 2.0 pretraining, a single iteration of self-training and decoding with a language model. Different to existing work, our approach does not leverage any other supervision than ST data. Code and models will be publicly released.

Changhan Wang, Anne Wu, Juan Pino, Alexei Baevski, Michael Auli, Alexis Conneau• 2021

Related benchmarks

TaskDatasetResultRank
Automatic Speech RecognitionLibriSpeech (test-other)
WER9.4
966
Automatic Speech RecognitionLibriSpeech clean (test)
WER5.6
833
Automatic Speech RecognitionLibriSpeech (dev-other)
WER9.2
411
Automatic Speech RecognitionLibriSpeech (dev-clean)
WER (%)5.7
319
Speech TranslationCoVoST-2 (test)
Avg BLEU (15 Dir)26.6
46
Spoken Named Entity RecognitionSLUE-VoxPopuli (dev)
F1 Score0.556
7
Spoken Named Entity RecognitionSLUE-VoxPopuli (test)
F1 Score50.9
6
Language IdentificationVoxLingua107 (dev)
Average Error Rate7.2
4
Showing 8 of 8 rows

Other info

Follow for update