Text-to-Text Pre-Training for Data-to-Text Tasks
About
We study the pre-train + fine-tune strategy for data-to-text tasks. Our experiments indicate that text-to-text pre-training in the form of T5, enables simple, end-to-end transformer based models to outperform pipelined neural architectures tailored for data-to-text generation, as well as alternative language model based pre-training techniques such as BERT and GPT-2. Importantly, T5 pre-training leads to better generalization, as evidenced by large improvements on out-of-domain test sets. We hope our work serves as a useful baseline for future research, as transfer learning becomes ever more prevalent for data-to-text tasks.
Mihir Kale, Abhinav Rastogi• 2020
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Data-to-text generation | WebNLG (test) | BLEU57.1 | 39 | |
| Response Generation | MultiWOZ (test) | BLEU Score35.1 | 27 | |
| Data-to-text generation | ToTTo | BLEU49.5 | 18 | |
| Table-to-text generation | Logic2Text (test) | BLEURT Score-1.079 | 18 | |
| Table-to-text generation | ToTTo (test) | BLEURT Score0.23 | 15 | |
| Table-to-text generation | Totto All (dev) | BLEURT0.233 | 15 | |
| Table-to-text generation | ToTTo Non (test) | BLEURT Score0.106 | 15 | |
| Table-to-text generation | ToTTo Over (test) | BLEURT0.354 | 15 | |
| Loosely controlled table-to-text generation | ToTTO Logic2Text-style (test) | BLEU29.4 | 15 | |
| Graph-to-text generation | WebNLG seen v1.0 (test) | BLEU63.9 | 12 |
Showing 10 of 36 rows