Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Non-Autoregressive Neural Machine Translation

About

Existing approaches to neural machine translation condition each output word on previously generated outputs. We introduce a model that avoids this autoregressive property and produces its outputs in parallel, allowing an order of magnitude lower latency during inference. Through knowledge distillation, the use of input token fertilities as a latent variable, and policy gradient fine-tuning, we achieve this at a cost of as little as 2.0 BLEU points relative to the autoregressive Transformer network used as a teacher. We demonstrate substantial cumulative improvements associated with each of the three aspects of our training strategy, and validate our approach on IWSLT 2016 English-German and two WMT language pairs. By sampling fertilities in parallel at inference time, our non-autoregressive model achieves near-state-of-the-art performance of 29.8 BLEU on WMT 2016 English-Romanian.

Jiatao Gu, James Bradbury, Caiming Xiong, Victor O.K. Li, Richard Socher• 2017

Related benchmarks

TaskDatasetResultRank
Machine TranslationWMT En-De 2014 (test)
BLEU21.47
379
Machine TranslationIWSLT De-En 2014 (test)
BLEU23.78
146
Machine TranslationWMT English-German 2014 (test)
BLEU22.7
136
Machine TranslationWMT 2014 (test)
BLEU22.42
100
Machine TranslationWMT En-De '14
BLEU19.17
89
Machine TranslationWMT Ro-En 2016 (test)
BLEU31.44
82
Machine TranslationWMT14 En-De newstest2014 (test)
BLEU20.36
65
Machine TranslationWMT De-En 14 (test)
BLEU24.81
59
Machine TranslationWMT 2016 (test)
BLEU31.44
58
Machine TranslationWMT16 EN-RO (test)
BLEU29.79
56
Showing 10 of 20 rows

Other info

Code

Follow for update