Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

AutoMixAlign: Adaptive Data Mixing for Multi-Task Preference Optimization in LLMs

About

When aligning large language models (LLMs), their performance on various tasks (such as being helpful, harmless, and honest) depends heavily on the composition of their training data. However, selecting a data mixture that achieves strong performance across all tasks is challenging. Existing approaches rely on large ablation studies, heuristics, or human intuition, but these can be prohibitively expensive and suboptimal. We study this problem in the setting of preference optimization via DPO and introduce AutoMixAlign (AMA), a theoretically-grounded algorithm that adaptively mixes datasets during training to balance performance across tasks. AMA first trains \textit{specialist models} for each task to determine losses that correspond to strong task performance. Then, it trains a generalist model using a novel minimax optimization that prioritizes tasks for which generalist model losses deviate most from specialist model losses. To optimize this problem, we propose two algorithms: (1) AMA-R, which adaptively reweights the objective to prioritize tasks, and (2) AMA-S, which adaptively adjusts how much data is sampled from each task to prioritize tasks. Both algorithms achieve a convergence rate of $O(1/\sqrt{T})$ in the convex case. AMA-R's convergence result follows from Sagawa et al. (2019), and we provide a convergence proof for AMA-S using online learning techniques such as EXP3. We evaluate AMA on several multitask alignment setups and find that AMA outperforms the standard alignment approach -- which simply optimizes the total loss across all tasks -- and also outperforms model merging methods.

Nicholas E. Corrado, Julian Katz-Samuels, Adithya Devraj, Hyokun Yun, Chao Zhang, Yi Xu, Yi Pan, Bing Yin, Trishul Chilimbi• 2025

Related benchmarks

TaskDatasetResultRank
Code GenerationHumanEval--
850
Instruction FollowingIFEval--
292
Code GenerationMBPP
Accuracy (%)55.76
146
Instruction FollowingAlpacaEval
Win Rate18.15
125
Instruction FollowingIFEval (test)
IFEval Score44.55
45
HelpfulnessAlpaca Eval
Alpaca Eval (%)17.77
22
Code GenerationMBPP
MBPP Accuracy51.44
22
HarmlessnessToxigen
Toxigen (%)99.99
17
LLM AlignmentCombined Suite Setup 3
Average Percentage Score54.38
9
Showing 9 of 9 rows

Other info

Follow for update