Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

MAUVE: Measuring the Gap Between Neural Text and Human Text using Divergence Frontiers

About

As major progress is made in open-ended text generation, measuring how close machine-generated text is to human language remains a critical open problem. We introduce MAUVE, a comparison measure for open-ended text generation, which directly compares the learnt distribution from a text generation model to the distribution of human-written text using divergence frontiers. MAUVE scales up to modern text generation models by computing information divergences in a quantized embedding space. Through an extensive empirical study on three open-ended generation tasks, we find that MAUVE identifies known properties of generated text, scales naturally with model size, and correlates with human judgments, with fewer restrictions than existing distributional evaluation metrics.

Krishna Pillutla, Swabha Swayamdipta, Rowan Zellers, John Thickstun, Sean Welleck, Yejin Choi, Zaid Harchaoui• 2021

Related benchmarks

TaskDatasetResultRank
Text GenerationWebText--
9
Human Correlation AnalysisOriginal human judgment dataset
Generation Perplexity0.643
3
Human Correlation AnalysisRefined human judgment dataset human vs model-generated--
3
Text Generation Evaluation CorrelationWebText (test)--
3
Open-ended text generation evaluation (Human-like)Web text--
2
Open-ended text generation evaluation (Discriminator Accuracy)News--
1
Open-ended text generation evaluation (Discriminator Accuracy)Stories--
1
Showing 7 of 7 rows

Other info

Code

Follow for update