Leveraging Pseudo-labeled Data to Improve Direct Speech-to-Speech Translation
About
Direct Speech-to-speech translation (S2ST) has drawn more and more attention recently. The task is very challenging due to data scarcity and complex speech-to-speech mapping. In this paper, we report our recent achievements in S2ST. Firstly, we build a S2ST Transformer baseline which outperforms the original Translatotron. Secondly, we utilize the external data by pseudo-labeling and obtain a new state-of-the-art result on the Fisher English-to-Spanish test set. Indeed, we exploit the pseudo data with a combination of popular techniques which are not trivial when applied to S2ST. Moreover, we evaluate our approach on both syntactically similar (Spanish-English) and distant (English-Chinese) language pairs. Our implementation is available at https://github.com/fengpeng-yue/speech-to-speech-translation.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Speech-to-speech translation | Fisher Spanish-English (test) | BLEU (Speech Input)46.3 | 55 | |
| Speech-to-speech translation | Fisher Spanish-English (dev) | BLEU (Speech)45.5 | 48 | |
| Speech-to-speech translation | CVSS-C | Avg Score0.273 | 38 | |
| Speech-to-speech translation | Fisher Spanish-English (dev2) | ASR BLEU47.6 | 36 | |
| Speech-to-speech translation | Fisher Es→En (dev) | ASR chrF63.8 | 10 | |
| Speech-to-speech translation | Fisher Es→En (test) | ASR chrF64.9 | 10 |