Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

PLOT: Enhancing Preference Learning via Optimal Transport

About

Preference learning in Large Language Models (LLMs) has advanced significantly, yet existing methods remain limited by modest performance gains, high computational costs, hyperparameter sensitivity, and insufficient modeling of global token-level relationships. We introduce PLOT, which enhances Preference Learning in fine-tuning-based alignment through a token-level loss derived from Optimal Transport. By formulating preference learning as an Optimal Transport Problem, PLOT aligns model outputs with human preferences while preserving the original distribution of LLMs, ensuring stability and robustness. Furthermore, PLOT leverages token embeddings to capture semantic relationships, enabling globally informed optimization. Experiments across two preference categories - Human Values and Logic & Problem Solving - spanning seven subpreferences demonstrate that PLOT consistently improves alignment performance while maintaining fluency and coherence. These results substantiate optimal transport as a principled methodology for preference learning, establishing a theoretically grounded framework that provides new insights for preference learning of LLMs.

Liang Zhu, Yuelin Bai, Xiankun Ren, Jiaxi Yang, Lei Zhang, Feiteng Fang, Hamid Alinejad-Rokny, Minghuan Tan, Min Yang• 2026

Related benchmarks

TaskDatasetResultRank
MathGSM8K
Accuracy0.6836
206
MathematicsMATH
MATH Accuracy48.28
85
Red Teaming AttackHarmBench (test)
ZS13.4
27
Logic & Problem SolvingMT-Bench
Reasoning Score4.5
3
Preference AlignmentHuman Values
Helpfulness (Reward)72.14
3
Showing 5 of 5 rows

Other info

Follow for update