Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Multilingual Pre-training with Language and Task Adaptation for Multilingual Text Style Transfer

About

We exploit the pre-trained seq2seq model mBART for multilingual text style transfer. Using machine translated data as well as gold aligned English sentences yields state-of-the-art results in the three target languages we consider. Besides, in view of the general scarcity of parallel data, we propose a modular approach for multilingual formality transfer, which consists of two training strategies that target adaptation to both language and task. Our approach achieves competitive performance without monolingual task-specific parallel data and can be applied to other style transfer tasks as well as to other languages.

Huiyuan Lai, Antonio Toral, Malvina Nissim• 2022

Related benchmarks

TaskDatasetResultRank
Informal to Formal Formality TransferPortuguese Informal to Formal (test)
BLEU55.3
13
Informal to Formal Formality TransferItalian Informal to Formal GYAFC equivalent (test)
BLEU48
13
Informal to Formal Formality TransferFrench Informal to Formal (test)
BLEU0.547
13
Formal to Informal Formality TransferItalian Formal to Informal (test)
BLEU0.186
10
Formal to Informal Formality TransferFrench (Formal to Informal) (test)
BLEU0.216
10
Formal to Informal Formality TransferPortuguese Formal to Informal (test)
BLEU0.225
10
Formal-to-Informal TransferItalian swapped setting (b) (test)
BLEU0.194
3
Formal-to-Informal TransferFrench swapped setting (b) (test)
BLEU20.1
3
Informal-to-Formal TransferItalian original setting (a) (test)
BLEU19.4
3
Informal-to-Formal TransferFrench setting (a) (test)
BLEU23.4
3
Showing 10 of 12 rows

Other info

Code

Follow for update