Hierarchical Transformers for Multi-Document Summarization
About
In this paper, we develop a neural summarization model which can effectively process multiple input documents and distill Transformer architecture with the ability to encode documents in a hierarchical manner. We represent cross-document relationships via an attention mechanism which allows to share information as opposed to simply concatenating text spans and processing them as a flat sequence. Our model learns latent dependencies among textual units, but can also take advantage of explicit graph representations focusing on similarity or discourse relations. Empirical results on the WikiSum dataset demonstrate that the proposed architecture brings substantial improvements over several strong baselines.
Yang Liu, Mirella Lapata• 2019
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Multi-document summarization | WikiSUM (test) | ROUGE-141.53 | 14 | |
| Abstractive Summarization | WikiCatSum Company (test) | Completeness2.96 | 5 | |
| Abstractive Summarization | WikiCatSum Film (test) | Completeness3.13 | 5 | |
| Abstractive Summarization | WikiCatSum Animal (test) | Completeness2.8 | 5 | |
| Wikipedia Abstract Generation | WikiCatSum Film (test) | ROUGE-124.6 | 5 | |
| Wikipedia Abstract Generation | WikiCatSum Company (test) | ROUGE-10.133 | 5 | |
| Wikipedia Abstract Generation | WikiCatSum Animal (test) | ROUGE-116.5 | 5 |
Showing 7 of 7 rows