Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Multi-Way, Multilingual Neural Machine Translation with a Shared Attention Mechanism

About

We propose multi-way, multilingual neural machine translation. The proposed approach enables a single neural translation model to translate between multiple languages, with a number of parameters that grows only linearly with the number of languages. This is made possible by having a single attention mechanism that is shared across all language pairs. We train the proposed multi-way, multilingual model on ten language pairs from WMT'15 simultaneously and observe clear performance improvements over models trained on only one language pair. In particular, we observe that the proposed model significantly improves the translation quality of low-resource language pairs.

Orhan Firat, Kyunghyun Cho, Yoshua Bengio• 2016

Related benchmarks

TaskDatasetResultRank
Document-Level Machine TranslationIWSLT Fr-En 2010 (test)
BLEU29.41
15
Document-Level Machine TranslationEuroparl WMT20 7 (test)
Human Rating3.31
8
Document-level Machine Translation (En -> Xx)Europarl-7 (test)
BLEU18.82
7
Document-level Machine Translation (En -> Xx)IWSLT-10 (test)
BLEU25.39
7
Document-level Machine Translation (Xx -> En)Europarl-7 (test)
BLEU22.4
7
Machine Translation (X→En)Flores-101 (test)
BLEU (low resource)8.8
6
Machine Translation (X→En)TED (test)
BLEU (Low Resource)0.199
6
Showing 7 of 7 rows

Other info

Follow for update