Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Small-Margin Preferences Still Matter-If You Train Them Right

About

Preference optimization methods such as DPO align large language models (LLMs) using paired comparisons, but their effectiveness can be highly sensitive to the quality and difficulty of preference pairs. A common heuristic treats small-margin (ambiguous) pairs as noisy and filters them out. In this paper, we revisit this assumption and show that pair difficulty interacts strongly with the optimization objective: when trained with preference-based losses, difficult pairs can destabilize training and harm alignment, yet these same pairs still contain useful supervision signals when optimized with supervised fine-tuning (SFT). Motivated by this observation, we propose MixDPO, a simple yet effective difficulty-aware training strategy that (i) orders preference data from easy to hard (a curriculum over margin-defined difficulty), and (ii) routes difficult pairs to an SFT objective while applying a preference loss to easy pairs. This hybrid design provides a practical mechanism to leverage ambiguous pairs without incurring the optimization failures often associated with preference losses on low-margin data. Across three LLM-judge benchmarks, MixDPO consistently improves alignment over DPO and a range of widely-used variants, with particularly strong gains on AlpacaEval~2 length-controlled (LC) win rate.

Jinlong Pang, Zhaowei Zhu, Na Di, Yichi Zhang, Yaxuan Wang, Chen Qian, Yang Liu• 2026

Related benchmarks

TaskDatasetResultRank
Instruction FollowingAlpacaEval 2.0
LC Win Rate29.47
281
Instruction FollowingAlpacaEval 2.0 (test)
LC Win Rate (%)29.47
71
LLM Alignment EvaluationAlpacaEval 2.0 (test)
LC Win Rate14.42
51
Downstream Task EvaluationOpenLLM Leaderboard v1 (test)
MMLU (5-shot)63.2
14
Preference EvaluationAlpacaEval 2
WR (%)559
12
Preference AlignmentArgilla-7k (test)
LC Win Rate9.23
5
Showing 6 of 6 rows

Other info

Follow for update