Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

A Novel Graph-based Multi-modal Fusion Encoder for Neural Machine Translation

About

Multi-modal neural machine translation (NMT) aims to translate source sentences into a target language paired with images. However, dominant multi-modal NMT models do not fully exploit fine-grained semantic correspondences between semantic units of different modalities, which have potential to refine multi-modal representation learning. To deal with this issue, in this paper, we propose a novel graph-based multi-modal fusion encoder for NMT. Specifically, we first represent the input sentence and image using a unified multi-modal graph, which captures various semantic relationships between multi-modal semantic units (words and visual objects). We then stack multiple graph-based multi-modal fusion layers that iteratively perform semantic interactions to learn node representations. Finally, these representations provide an attention-based context vector for the decoder. We evaluate our proposed encoder on the Multi30K datasets. Experimental results and in-depth analysis show the superiority of our multi-modal NMT model.

Yongjing Yin, Fandong Meng, Jinsong Su, Chulun Zhou, Zhengyuan Yang, Jie Zhou, Jiebo Luo• 2020

Related benchmarks

TaskDatasetResultRank
Multimodal Machine TranslationMulti30K (test)--
139
Multimodal Machine Translation (English-German)Multi30K 2016 (test)
BLEU39.8
52
Multimodal Machine TranslationMulti30k En-De 2017 (test)
METEOR51.9
45
Multimodal Machine TranslationMulti30k En-Fr 2017 (test)
METEOR69.3
31
Machine TranslationMulti30k En→Fr v1 2017 (test)
BLEU53.9
30
Multimodal Machine TranslationMulti30k En-Fr 2016 (test)
METEOR Score74.9
30
Machine TranslationMulti30k Task1 (en-de)
BLEU Score39.8
26
Machine TranslationMulti30K En → De (test)
METEOR45.7
26
Machine TranslationMulti30k Task1 en-fr
BLEU Score60.9
25
Machine TranslationMulti30k M30kT (test)
BLEU Score32.2
19
Showing 10 of 26 rows

Other info

Follow for update