Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Aligning Large Language Models with Implicit Preferences from User-Generated Content

About

Learning from preference feedback is essential for aligning large language models (LLMs) with human values and improving the quality of generated responses. However, existing preference learning methods rely heavily on curated data from humans or advanced LLMs, which is costly and difficult to scale. In this work, we present PUGC, a novel framework that leverages implicit human Preferences in unlabeled User-Generated Content (UGC) to generate preference data. Although UGC is not explicitly created to guide LLMs in generating human-preferred responses, it often reflects valuable insights and implicit preferences from its creators that has the potential to address readers' questions. PUGC transforms UGC into user queries and generates responses from the policy model. The UGC is then leveraged as a reference text for response scoring, aligning the model with these implicit preferences. This approach improves the quality of preference data while enabling scalable, domain-specific alignment. Experimental results on Alpaca Eval 2 show that models trained with DPO and PUGC achieve a 9.37% performance improvement over traditional methods, setting a 35.93% state-of-the-art length-controlled win rate using Mistral-7B-Instruct. Further studies highlight gains in reward quality, domain-specific alignment effectiveness, robustness against UGC quality, and theory of mind capabilities. Our code and dataset are available at https://zhaoxuan.info/PUGC.github.io/

Zhaoxuan Tan, Zheng Li, Tianyi Liu, Haodong Wang, Hyokun Yun, Ming Zeng, Pei Chen, Zhihan Zhang, Yifan Gao, Ruijie Wang, Priyanka Nigam, Bing Yin, Meng Jiang• 2025

Related benchmarks

TaskDatasetResultRank
Multi-turn Dialogue EvaluationMT-Bench--
331
Instruction FollowingIFEval--
292
Instruction FollowingAlpacaEval 2.0
LC Win Rate35.93
281
Mathematical ReasoningGSM8K
Accuracy41.17
212
Multitask Language UnderstandingMMLU-Pro
Accuracy28.37
99
LLM Alignment EvaluationArena Hard
Win Rate13.7
67
Safety EvaluationSafetyBench en
Avg Score72
25
Theory of Mind reasoningBigTOM (All)
Accuracy84.4
24
Question AnsweringTruthfulQA
Accuracy42.77
7
General Language Model EvaluationWildBench
WildBench Score26.95
2
Showing 10 of 10 rows

Other info

Code

Follow for update