Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Neural Latent Extractive Document Summarization

About

Extractive summarization models require sentence-level labels, which are usually created heuristically (e.g., with rule-based methods) given that most summarization datasets only have document-summary pairs. Since these labels might be suboptimal, we propose a latent variable extractive model where sentences are viewed as latent variables and sentences with activated variables are used to infer gold summaries. During training the loss comes \emph{directly} from gold summaries. Experiments on the CNN/Dailymail dataset show that our model improves over a strong extractive baseline trained on heuristically approximated labels and also performs competitively to several recent models.

Xingxing Zhang, Mirella Lapata, Furu Wei, Ming Zhou• 2018

Related benchmarks

TaskDatasetResultRank
SummarizationCNN/Daily Mail original, non-anonymized (test)
ROUGE-141.05
54
Abstractive SummarizationCNN/DailyMail full length F-1 (test)
ROUGE-141.05
48
Extractive SummarizationCNN/Daily Mail (test)
ROUGE-15
36
SummarizationCNNDM full-length F1 (test)
ROUGE-141.05
19
Showing 4 of 4 rows

Other info

Follow for update