Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Neural Extractive Text Summarization with Syntactic Compression

About

Recent neural network approaches to summarization are largely either selection-based extraction or generation-based abstraction. In this work, we present a neural model for single-document summarization based on joint extraction and syntactic compression. Our model chooses sentences from the document, identifies possible compressions based on constituency parses, and scores those compressions with a neural model to produce the final summary. For learning, we construct oracle extractive-compressive summaries, then learn both of our components jointly with this supervision. Experimental results on the CNN/Daily Mail and New York Times datasets show that our model achieves strong performance (comparable to state-of-the-art systems) as evaluated by ROUGE. Moreover, our approach outperforms an off-the-shelf compression module, and human and manual evaluation shows that our model's output generally remains grammatical.

Jiacheng Xu, Greg Durrett• 2019

Related benchmarks

TaskDatasetResultRank
Extractive SummarizationNYT50 (test)
ROUGE-145.5
21
SummarizationCNNDM full-length F1 (test)
ROUGE-141.7
19
SummarizationCNN/Daily Mail full length (test)
ROUGE-141.7
18
Extractive SummarizationCNN-DM (test)
ROUGE-141.7
18
Showing 4 of 4 rows

Other info

Follow for update