Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Generating Wikipedia by Summarizing Long Sequences

About

We show that generating English Wikipedia articles can be approached as a multi- document summarization of source documents. We use extractive summarization to coarsely identify salient information and a neural abstractive model to generate the article. For the abstractive model, we introduce a decoder-only architecture that can scalably attend to very long sequences, much longer than typical encoder- decoder architectures used in sequence transduction. We show that this model can generate fluent, coherent multi-sentence paragraphs and even whole Wikipedia articles. When given reference documents, we show it can extract relevant factual information as reflected in perplexity, ROUGE scores and human evaluations.

Peter J. Liu, Mohammad Saleh, Etienne Pot, Ben Goodrich, Ryan Sepassi, Lukasz Kaiser, Noam Shazeer• 2018

Related benchmarks

TaskDatasetResultRank
Multi-document summarizationWikiSUM (test)
ROUGE-140.77
14
Lead Paragraph GenerationWikiSum CommonCrawl static (test)
ROUGE-L38.8
8
System Paradigm ComparisonKnowledge Materialization and Encyclopedia Generation Paradigms
Scale10
8
Music ModelingPiano-e-Competition (val)
Val NLL1.863
6
Abstractive SummarizationWikiCatSum Company (test)
Completeness2.69
5
Wikipedia Abstract GenerationWikiCatSum Company (test)
ROUGE-10.197
5
Wikipedia Abstract GenerationWikiCatSum Animal (test)
ROUGE-125.2
5
Abstractive SummarizationWikiCatSum Film (test)
Completeness1.93
5
Abstractive SummarizationWikiCatSum Animal (test)
Completeness2.22
5
Wikipedia Abstract GenerationWikiCatSum Film (test)
ROUGE-119.8
5
Showing 10 of 10 rows

Other info

Follow for update