ENGINE: Energy-Based Inference Networks for Non-Autoregressive Machine Translation
About
We propose to train a non-autoregressive machine translation model to minimize the energy defined by a pretrained autoregressive model. In particular, we view our non-autoregressive translation system as an inference network (Tu and Gimpel, 2018) trained to minimize the autoregressive teacher energy. This contrasts with the popular approach of training a non-autoregressive model on a distilled corpus consisting of the beam-searched outputs of such a teacher model. Our approach, which we call ENGINE (ENerGy-based Inference NEtworks), achieves state-of-the-art non-autoregressive results on the IWSLT 2014 DE-EN and WMT 2016 RO-EN datasets, approaching the performance of autoregressive models.
Lifu Tu, Richard Yuanzhe Pang, Sam Wiseman, Kevin Gimpel• 2020
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Machine Translation | WMT De-En 14 (test) | BLEU28.1 | 59 | |
| Table-to-text generation | DART | METEOR0.3651 | 30 | |
| Machine Translation | WMT16 Ro-En (test) | BLEU28.2 | 27 | |
| Abstractive Summarization | Xsum | -- | 14 | |
| Dialogue Generation | Commonsense Dialogue (CD) | BLEU-110.13 | 9 |
Showing 5 of 5 rows