Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Verbalized Sampling: How to Mitigate Mode Collapse and Unlock LLM Diversity

About

Post-training alignment often reduces LLM diversity, leading to a phenomenon known as mode collapse. Unlike prior work that attributes this effect to algorithmic limitations, we identify a fundamental, pervasive data-level driver: typicality bias in preference data, whereby annotators systematically favor familiar text as a result of well-established findings in cognitive psychology. We formalize this bias theoretically, verify it on preference datasets empirically, and show that it plays a central role in mode collapse. Motivated by this analysis, we introduce Verbalized Sampling, a simple, training-free prompting strategy to circumvent mode collapse. VS prompts the model to verbalize a probability distribution over a set of responses (e.g., "Generate 5 jokes about coffee and their corresponding probabilities"). Comprehensive experiments show that VS significantly improves performance across creative writing (poems, stories, jokes), dialogue simulation, open-ended QA, and synthetic data generation, without sacrificing factual accuracy and safety. For instance, in creative writing, VS increases diversity by 1.6-2.1x over direct prompting. We further observe an emergent trend that more capable models benefit more from VS. In sum, our work provides a new data-centric perspective on mode collapse and a practical inference-time remedy that helps unlock pre-trained generative diversity.

Jiayi Zhang, Simon Yu, Derek Chong, Anthony Sicilia, Michael R. Tomz, Christopher D. Manning, Weiyan Shi• 2025

Related benchmarks

TaskDatasetResultRank
Joke generationJoke
Quality0.8423
29
Empathic DialogueLend-an-Ear (test)
Average Word Count88.1
23
Creative WritingStory
Semantic Diversity38.6
20
Creative WritingOverall Average Poem, Joke, Story
Semantic Diversity0.3603
20
Creative WritingPoem
Semantic Diversity30.95
20
Open-ended Information SeekingInfinity-Chat (test)
DSem0.417
20
Downstream ClassificationEducation Unconstrained
F1 Score7.2
4
Downstream ClassificationWriting Unconstrained
F1 Score22.1
3
Downstream ClassificationEducation Category-controlled top-K
F1 Score43.5
3
Downstream ClassificationPackaging Category-controlled top-K
F1 Score15.8
3
Showing 10 of 15 rows

Other info

GitHub

Follow for update