Few-Shot Data-to-Text Generation via Unified Representation and Multi-Source Learning
About
We present a novel approach for structured data-to-text generation that addresses the limitations of existing methods that primarily focus on specific types of structured data. Our proposed method aims to improve performance in multi-task training, zero-shot and few-shot scenarios by providing a unified representation that can handle various forms of structured data such as tables, knowledge graph triples, and meaning representations. We demonstrate that our proposed approach can effectively adapt to new structured forms, and can improve performance in comparison to current methods. For example, our method resulted in a 66% improvement in zero-shot BLEU scores when transferring models trained on table inputs to a knowledge graph dataset. Our proposed method is an important step towards a more general data-to-text generation framework.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Data-to-text generation | DART (test) | BLEU50.2 | 42 | |
| Data-to-text generation | E2E (test) | BLEU43.2 | 33 | |
| Data-to-text generation | ToTTo full (test) | BLEU50.8 | 12 | |
| Logical Data-to-Text Generation | LOGICNLG (test) | BLEU-325.4 | 10 | |
| Data-to-text generation | DART KG | BLEU0.315 | 5 | |
| Data-to-text generation | WebNLG KG | BLEU (Unigram)39.8 | 5 | |
| Data-to-text generation | E2E clean MR | BLEU22.6 | 4 | |
| Table-to-text generation | LogicNLG Table | BLEU-38.9 | 3 |