Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Massively Multilingual Neural Machine Translation

About

Multilingual neural machine translation (NMT) enables training a single model that supports translation from multiple source languages into multiple target languages. In this paper, we push the limits of multilingual NMT in terms of number of languages being used. We perform extensive experiments in training massively multilingual NMT models, translating up to 102 languages to and from English within a single model. We explore different setups for training such models and analyze the trade-offs between translation quality and various modeling decisions. We report results on the publicly available TED talks multilingual corpus where we show that massively multilingual many-to-many models are effective in low resource settings, outperforming the previous state-of-the-art while supporting up to 59 languages. Our experiments on a large-scale dataset with 102 languages to and from English and up to one million examples per direction also show promising results, surpassing strong bilingual baselines and encouraging future work on massively multilingual NMT.

Roee Aharoni, Melvin Johnson, Orhan Firat• 2019

Related benchmarks

TaskDatasetResultRank
Machine TranslationOPUS-7 (test)
Translation Score (X -> Ar)21.61
17
Multilingual Machine TranslationOPUS-7
Ar Score79.03
16
Machine Translationsk-en (test)
BLEU29.54
15
Machine TranslationTed Talk aze-eng (test)
BLEU12.8
9
Machine TranslationTed Talk bel-eng (test)
BLEU21.7
9
Machine TranslationTed Talk glg-eng (test)
BLEU30.7
9
Machine TranslationTed Talk slk-eng (test)
BLEU29.5
9
Machine TranslationTed Talk Average (test)
BLEU Score28.04
9
Multilingual Many-to-One Machine TranslationTED-59
AZ Score11.24
9
Machine TranslationTED low-resource En-Sk (test)
BLEU24.52
7
Showing 10 of 19 rows

Other info

Follow for update