Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

A Well-Composed Text is Half Done! Composition Sampling for Diverse Conditional Generation

About

We propose Composition Sampling, a simple but effective method to generate diverse outputs for conditional generation of higher quality compared to previous stochastic decoding strategies. It builds on recently proposed plan-based neural generation models (Narayan et al, 2021) that are trained to first create a composition of the output and then generate by conditioning on it and the input. Our approach avoids text degeneration by first sampling a composition in the form of an entity chain and then using beam search to generate the best possible text grounded to this entity chain. Experiments on summarization (CNN/DailyMail and XSum) and question generation (SQuAD), using existing and newly proposed automatic metrics together with human-based evaluation, demonstrate that Composition Sampling is currently the best available decoding strategy for generating diverse meaningful outputs.

Shashi Narayan, Gon\c{c}alo Sim\~oes, Yao Zhao, Joshua Maynez, Dipanjan Das, Michael Collins, Mirella Lapata• 2022

Related benchmarks

TaskDatasetResultRank
Multi-hop Question Answering2WikiMultihopQA
EM60.5
387
SummarizationXSum (test)
ROUGE-225.06
246
Question GenerationSQuAD (test)
BLEU-121.04
22
Multi-hop Question AnsweringHotpotQA
EM55
10
SummarizationXSum 50 document sample (sampled)
RL Score40.9
9
SummarizationCNN/DailyMail 50 document sample (sampled)
PPL0.3
7
Multi-hop Question GenerationHotpotQA
Pairwise BLEU89.5
4
Showing 7 of 7 rows

Other info

Code

Follow for update