Duplex Sequence-to-Sequence Learning for Reversible Machine Translation
About
Sequence-to-sequence learning naturally has two directions. How to effectively utilize supervision signals from both directions? Existing approaches either require two separate models, or a multitask-learned model but with inferior performance. In this paper, we propose REDER (Reversible Duplex Transformer), a parameter-efficient model and apply it to machine translation. Either end of REDER can simultaneously input and output a distinct language. Thus REDER enables reversible machine translation by simply flipping the input and output ends. Experiments verify that REDER achieves the first success of reversible machine translation, which helps outperform its multitask-trained baselines by up to 1.3 BLEU.
Zaixiang Zheng, Hao Zhou, Shujian Huang, Jiajun Chen, Jingjing Xu, Lei Li• 2021
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Machine Translation | WMT En-De 2014 (test) | BLEU27.5 | 379 | |
| Machine Translation | WMT Ro-En 2016 (test) | BLEU34.03 | 82 | |
| Machine Translation | WMT16 EN-RO (test) | BLEU33.6 | 56 | |
| Machine Translation | WMT14 DE-EN (test) | BLEU31.25 | 28 | |
| Machine Translation | WMT20 JA-EN (test) | BLEU20.7 | 8 | |
| Machine Translation | WMT20 EN-JA (test) | BLEU20 | 2 |
Showing 6 of 6 rows