Thank you BART! Rewarding Pre-Trained Models Improves Formality Style Transfer
About
Scarcity of parallel data causes formality style transfer models to have scarce success in preserving content. We show that fine-tuning pre-trained language (GPT-2) and sequence-to-sequence (BART) models boosts content preservation, and that this is possible even with limited amounts of parallel data. Augmenting these models with rewards that target style and content -- the two core aspects of the task -- we achieve a new state-of-the-art.
Huiyuan Lai, Antonio Toral, Malvina Nissim• 2021
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Formality Style Transfer | GYAFC Entertainment & Music 1.0 (test) | BLEURT0.274 | 15 | |
| Formality Style Transfer | GYAFC Family & Relationships 1.0 (test) | BLEU0.793 | 15 | |
| Formality Style Transfer | GYAFC Entertainment & Music (test) | BLEU76.5 | 10 | |
| Formality Style Transfer | GYAFC Family & Relationships (test) | BLEU79.25 | 10 |
Showing 4 of 4 rows