ERNIE-GEN: An Enhanced Multi-Flow Pre-training and Fine-tuning Framework for Natural Language Generation
About
Current pre-training works in natural language generation pay little attention to the problem of exposure bias on downstream tasks. To address this issue, we propose an enhanced multi-flow sequence to sequence pre-training and fine-tuning framework named ERNIE-GEN, which bridges the discrepancy between training and inference with an infilling generation mechanism and a noise-aware generation method. To make generation closer to human writing patterns, this framework introduces a span-by-span generation flow that trains the model to predict semantically-complete spans consecutively rather than predicting word by word. Unlike existing pre-training methods, ERNIE-GEN incorporates multi-granularity target sampling to construct pre-training data, which enhances the correlation between encoder and decoder. Experimental results demonstrate that ERNIE-GEN achieves state-of-the-art results with a much smaller amount of pre-training data and parameters on a range of language generation tasks, including abstractive summarization (Gigaword and CNN/DailyMail), question generation (SQuAD), dialogue generation (Persona-Chat) and generative question answering (CoQA).
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Abstractive Text Summarization | CNN/Daily Mail (test) | ROUGE-L41.6 | 169 | |
| Abstractive Summarization | Gigaword (test) | ROUGE-139.46 | 58 | |
| Abstractive Summarization | CNN/DailyMail | ROUGE-144.02 | 25 | |
| Question Generation | SQuAD 1.1 | METEOR0.2631 | 21 | |
| Question Generation | SQuAD | BLEU-40.254 | 21 | |
| Dialogue Response Generation | Persona-Chat | BLEU-146.8 | 20 | |
| Question Generation | SQuAD 1.1 (dev) | BLEU-425.4 | 16 | |
| Conversational Question Answering | CoQA (dev) | Overall F10.845 | 14 | |
| Question Generation | SQuAD Du | BLEU-422.28 | 10 | |
| Question Generation | SQuAD 1.1 (reversed dev-test) | BLEU-426.95 | 9 |