Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Data Diversification: A Simple Strategy For Neural Machine Translation

About

We introduce Data Diversification: a simple but effective strategy to boost neural machine translation (NMT) performance. It diversifies the training data by using the predictions of multiple forward and backward models and then merging them with the original dataset on which the final NMT model is trained. Our method is applicable to all NMT models. It does not require extra monolingual data like back-translation, nor does it add more computations and parameters like ensembles of models. Our method achieves state-of-the-art BLEU scores of 30.7 and 43.7 in the WMT'14 English-German and English-French translation tasks, respectively. It also substantially improves on 8 other translation tasks: 4 IWSLT tasks (English-German and English-French) and 4 low-resource translation tasks (English-Nepali and English-Sinhala). We demonstrate that our method is more effective than knowledge distillation and dual learning, it exhibits strong correlation with ensembles of models, and it trades perplexity off for better BLEU score. We have released our source code at https://github.com/nxphi47/data_diversification

Xuan-Phi Nguyen, Shafiq Joty, Wu Kui, Ai Ti Aw• 2019

Related benchmarks

TaskDatasetResultRank
Machine TranslationWMT En-De 2014 (test)
BLEU27.9
379
Machine TranslationIWSLT De-En 2014 (test)
BLEU37
146
Machine TranslationWMT English-German 2014 (test)
BLEU30.7
136
Machine TranslationIWSLT German-to-English '14 (test)
BLEU Score37.2
110
Machine TranslationIWSLT En-De 2014 (test)
BLEU30.47
92
Machine TranslationWMT English-French 2014 (test)
BLEU43.7
41
Machine TranslationWMT14 English-French (newstest2014)
BLEU43.7
39
Machine TranslationIWSLT De-En 14
BLEU Score37.01
33
Machine TranslationIWSLT17 En-Fr (test)
BLEU40.67
18
Machine TranslationIWSLT De-En combined 2014 (test)
BLEU Score37.2
16
Showing 10 of 18 rows

Other info

Code

Follow for update