Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Integrating Vectorized Lexical Constraints for Neural Machine Translation

About

Lexically constrained neural machine translation (NMT), which controls the generation of NMT models with pre-specified constraints, is important in many practical scenarios. Due to the representation gap between discrete constraints and continuous vectors in NMT models, most existing works choose to construct synthetic data or modify the decoding algorithm to impose lexical constraints, treating the NMT model as a black box. In this work, we propose to open this black box by directly integrating the constraints into NMT models. Specifically, we vectorize source and target constraints into continuous keys and values, which can be utilized by the attention modules of NMT models. The proposed integration method is based on the assumption that the correspondence between keys and values in attention modules is naturally suitable for modeling constraint pairs. Experimental results show that our method consistently outperforms several representative baselines on four language pairs, demonstrating the superiority of integrating vectorized lexical constraints.

Shuo Wang, Zhixing Tan, Yang Liu• 2022

Related benchmarks

TaskDatasetResultRank
Entity TranslationWMT newstest (test)
Error Rate28.1
32
Machine TranslationDe-En (test)
BLEU35.9
23
Machine TranslationAmbiguous Constraint En-Zh 1.0 (test)
SacreBLEU30.1
10
Machine TranslationAmbiguous Constraint En-Zh (test)
SacreBLEU30.1
10
Machine TranslationAmbiguous Constraint De-En 1.0 (test)
SacreBLEU31.4
10
Machine TranslationAmbiguous Constraint De-En (test)
SacreBLEU31.4
10
Code-switched TranslationZh -> En code-switched (test)
BLEU33.5
6
Code-switched TranslationEn -> Zh code-switched (test)
BLEU56.7
6
Code-switched TranslationDe -> En code-switched (test)
BLEU32.9
6
Code-switched TranslationEn -> De code-switched (test)
BLEU27.1
6
Showing 10 of 16 rows

Other info

Code

Follow for update