Massively Multilingual Neural Machine Translation
About
Multilingual neural machine translation (NMT) enables training a single model that supports translation from multiple source languages into multiple target languages. In this paper, we push the limits of multilingual NMT in terms of number of languages being used. We perform extensive experiments in training massively multilingual NMT models, translating up to 102 languages to and from English within a single model. We explore different setups for training such models and analyze the trade-offs between translation quality and various modeling decisions. We report results on the publicly available TED talks multilingual corpus where we show that massively multilingual many-to-many models are effective in low resource settings, outperforming the previous state-of-the-art while supporting up to 59 languages. Our experiments on a large-scale dataset with 102 languages to and from English and up to one million examples per direction also show promising results, surpassing strong bilingual baselines and encouraging future work on massively multilingual NMT.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Machine Translation | OPUS-7 (test) | Translation Score (X -> Ar)21.61 | 17 | |
| Multilingual Machine Translation | OPUS-7 | Ar Score79.03 | 16 | |
| Machine Translation | sk-en (test) | BLEU29.54 | 15 | |
| Machine Translation | Ted Talk aze-eng (test) | BLEU12.8 | 9 | |
| Machine Translation | Ted Talk bel-eng (test) | BLEU21.7 | 9 | |
| Machine Translation | Ted Talk glg-eng (test) | BLEU30.7 | 9 | |
| Machine Translation | Ted Talk slk-eng (test) | BLEU29.5 | 9 | |
| Machine Translation | Ted Talk Average (test) | BLEU Score28.04 | 9 | |
| Multilingual Many-to-One Machine Translation | TED-59 | AZ Score11.24 | 9 | |
| Machine Translation | TED low-resource En-Sk (test) | BLEU24.52 | 7 |