Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Fast Abstractive Summarization with Reinforce-Selected Sentence Rewriting

About

Inspired by how humans summarize long documents, we propose an accurate and fast summarization model that first selects salient sentences and then rewrites them abstractively (i.e., compresses and paraphrases) to generate a concise overall summary. We use a novel sentence-level policy gradient method to bridge the non-differentiable computation between these two neural networks in a hierarchical way, while maintaining language fluency. Empirically, we achieve the new state-of-the-art on all metrics (including human evaluation) on the CNN/Daily Mail dataset, as well as significantly higher abstractiveness scores. Moreover, by first operating at the sentence-level and then the word-level, we enable parallel decoding of our neural generative model that results in substantially faster (10-20x) inference speed as well as 4x faster training convergence than previous long-paragraph encoder-decoder models. We also demonstrate the generalization of our model on the test-only DUC-2002 dataset, where we achieve higher scores than a state-of-the-art model.

Yen-Chun Chen, Mohit Bansal• 2018

Related benchmarks

TaskDatasetResultRank
Dialogue SummarizationSamSum (test)
ROUGE-218.1
80
Text SummarizationCNN/Daily Mail (test)
ROUGE-218.72
65
SummarizationbigPatent
ROUGE-137.12
61
SummarizationCNN/Daily Mail original, non-anonymized (test)
ROUGE-141.47
54
Abstractive SummarizationCNN/Daily Mail non-anonymous (test)
ROUGE-140.88
52
SummarizationCNNDM full-length F1 (test)
ROUGE-140.68
19
SummarizationEMAILSUM short 1.0 (test)
R131.15
19
SummarizationEMAILSUM long 1.0 (test)
ROUGE-1 (R1)39.35
19
SummarizationDUC 2002 (test)
ROUGE-139.46
18
SummarizationCNN/Daily Mail full length (test)
ROUGE-140.88
18
Showing 10 of 15 rows

Other info

Code

Follow for update