Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

ERNIE-GEN: An Enhanced Multi-Flow Pre-training and Fine-tuning Framework for Natural Language Generation

About

Current pre-training works in natural language generation pay little attention to the problem of exposure bias on downstream tasks. To address this issue, we propose an enhanced multi-flow sequence to sequence pre-training and fine-tuning framework named ERNIE-GEN, which bridges the discrepancy between training and inference with an infilling generation mechanism and a noise-aware generation method. To make generation closer to human writing patterns, this framework introduces a span-by-span generation flow that trains the model to predict semantically-complete spans consecutively rather than predicting word by word. Unlike existing pre-training methods, ERNIE-GEN incorporates multi-granularity target sampling to construct pre-training data, which enhances the correlation between encoder and decoder. Experimental results demonstrate that ERNIE-GEN achieves state-of-the-art results with a much smaller amount of pre-training data and parameters on a range of language generation tasks, including abstractive summarization (Gigaword and CNN/DailyMail), question generation (SQuAD), dialogue generation (Persona-Chat) and generative question answering (CoQA).

Dongling Xiao, Han Zhang, Yukun Li, Yu Sun, Hao Tian, Hua Wu, Haifeng Wang• 2020

Related benchmarks

TaskDatasetResultRank
Abstractive Text SummarizationCNN/Daily Mail (test)
ROUGE-L41.6
169
Abstractive SummarizationGigaword (test)
ROUGE-139.46
58
Abstractive SummarizationCNN/DailyMail
ROUGE-144.02
25
Question GenerationSQuAD 1.1
METEOR0.2631
21
Question GenerationSQuAD
BLEU-40.254
21
Dialogue Response GenerationPersona-Chat
BLEU-146.8
20
Question GenerationSQuAD 1.1 (dev)
BLEU-425.4
16
Conversational Question AnsweringCoQA (dev)
Overall F10.845
14
Question GenerationSQuAD Du
BLEU-422.28
10
Question GenerationSQuAD 1.1 (reversed dev-test)
BLEU-426.95
9
Showing 10 of 13 rows

Other info

Code

Follow for update