A Neural Attention Model for Abstractive Sentence Summarization
About
Summarization based on text extraction is inherently limited, but generation-style abstractive methods have proven challenging to build. In this work, we propose a fully data-driven approach to abstractive sentence summarization. Our method utilizes a local attention-based model that generates each word of the summary conditioned on the input sentence. While the model is structurally simple, it can easily be trained end-to-end and scales to a large amount of training data. The model shows significant performance gains on the DUC-2004 shared task compared with several strong baselines.
Alexander M. Rush, Sumit Chopra, Jason Weston• 2015
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Text Summarization | DUC 2004 (test) | ROUGE-128.18 | 115 | |
| Text Summarization | Gigaword (test) | ROUGE-137.41 | 75 | |
| Abstractive Summarization | Gigaword (test) | ROUGE-129.78 | 58 | |
| Summarization | Gigaword | ROUGE-L26.96 | 38 | |
| Abstractive Summarization | Gigawords (test) | ROUGE-132.7 | 27 | |
| Summarization | Summarization dataset | ROUGE-L F151.5 | 16 | |
| Abstractive Text Summarization | Gigaword | ROUGE-129.76 | 14 | |
| Code-to-NL generation | CodeNN C# (test) | BLEU19.31 | 13 | |
| Abstractive Summarization | Gigaword full-length F1 (test) | ROUGE-1 F129.78 | 12 | |
| Summarization | DUC 75-byte limit 2004 | ROUGE-1 Recall26.55 | 6 |
Showing 10 of 14 rows