SummaRuNNer: A Recurrent Neural Network based Sequence Model for Extractive Summarization of Documents
About
We present SummaRuNNer, a Recurrent Neural Network (RNN) based sequence model for extractive summarization of documents and show that it achieves performance better than or comparable to state-of-the-art. Our model has the additional advantage of being very interpretable, since it allows visualization of its predictions broken up by abstract features such as information content, salience and novelty. Another novel contribution of our work is abstractive training of our extractive model that can train on human generated reference summaries alone, eliminating the need for sentence-level extractive labels.
Ramesh Nallapati, Feifei Zhai, Bowen Zhou• 2016
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Summarization | XSum (test) | ROUGE-28.81 | 231 | |
| Abstractive Text Summarization | CNN/Daily Mail (test) | ROUGE-L36.67 | 169 | |
| Summarization | arXiv (test) | ROUGE-160 | 161 | |
| Summarization | PubMed (test) | ROUGE-161.99 | 107 | |
| Text Summarization | CNN/Daily Mail (test) | ROUGE-216.2 | 65 | |
| Summarization | CNN/Daily Mail original, non-anonymized (test) | ROUGE-139.6 | 54 | |
| Document Summarization | GovReport (test) | ROUGE-175.56 | 50 | |
| Abstractive Summarization | CNN/DailyMail full length F-1 (test) | ROUGE-139.6 | 48 | |
| Extractive Summarization | CNN/Daily Mail (test) | ROUGE-142 | 36 | |
| Summarization | CNNDM full-length F1 (test) | ROUGE-139.6 | 19 |
Showing 10 of 27 rows