Contrastive Learning for Many-to-many Multilingual Neural Machine Translation
About
Existing multilingual machine translation approaches mainly focus on English-centric directions, while the non-English directions still lag behind. In this work, we aim to build a many-to-many translation system with an emphasis on the quality of non-English language directions. Our intuition is based on the hypothesis that a universal cross-language representation leads to better multilingual translation performance. To this end, we propose mRASP2, a training method to obtain a single unified multilingual translation model. mRASP2 is empowered by two techniques: a) a contrastive learning scheme to close the gap among representations of different languages, and b) data augmentation on both multiple parallel and monolingual data to further align token representations. For English-centric directions, mRASP2 outperforms existing best unified model and achieves competitive or even better performance than the pre-trained and fine-tuned model mBART on tens of WMT's translation directions. For non-English directions, mRASP2 achieves an improvement of average 10+ BLEU compared with the multilingual Transformer baseline. Code, data and trained models are available at https://github.com/PANXiao1994/mRASP2.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Machine Translation | WMT En-Fr 2014 (test) | BLEU43.5 | 237 | |
| Machine Translation | WMT16 EN-RO (test) | BLEU39.1 | 56 | |
| Machine Translation | OPUS-100 (test) | Average BLEU Score15.31 | 19 | |
| Machine Translation | WMT En-Tr 17 | BLEU25.8 | 17 | |
| Machine Translation | OPUS-7 (test) | Translation Score (X -> Ar)74.36 | 17 | |
| Machine Translation | IWSLT 2017 (test) | De-It Translation Score77.01 | 15 | |
| Machine Translation | WMT En-Fi 17 (test) | BLEU (tokenized)30.1 | 14 | |
| Machine Translation | PC-6 (test) | Translation Score (x -> Cs)68.98 | 13 | |
| Machine Translation | WMT En-Es 13 (test) | Tokenized BLEU35 | 10 | |
| Machine Translation | WMT En-Tr 17 (test) | BLEU (tokenized)21.4 | 6 |