Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

ScaleBiO: Scalable Bilevel Optimization for LLM Data Reweighting

About

Bilevel optimization has shown its utility across various machine learning settings, yet most algorithms in practice require second-order information, making it challenging to scale them up. Only recently, a paradigm of first-order algorithms has emerged in the theoretical literature, capable of effectively addressing bilevel optimization problems. Nevertheless, the practical efficiency of this paradigm remains unverified, particularly in the context of large language models (LLMs). This paper introduces the first scalable instantiation of this paradigm called ScaleBiO, focusing on bilevel optimization for large-scale LLM data reweighting. By combining with a recently proposed memory-efficient training technique called LISA, our novel algorithm allows the paradigm to scale to $\sim$30B-sized LLMs on $8\times$H100 GPUs, marking the first successful application of bilevel optimization under practical scenarios for large-sized LLMs. Empirically, extensive experiments on data reweighting verify the effectiveness of ScaleBiO for different-scaled models, including Llama-3-8B, Gemma-2-9B, Qwen-2-7B, and Qwen-2.5-32B, where bilevel optimization succeeds in instruction-following and math reasoning tasks, outperforming several popular baselines, including uniform sampling, influence-aware data filtering, and reference-model-based sampling methods. Theoretically, ScaleBiO ensures the optimality of the learned data weights, along with a convergence guarantee matching the conventional first-order bilevel optimization paradigm on smooth and strongly convex objectives.

Rui Pan, Dylan Zhang, Hanning Zhang, Xingyuan Pan, Minrui Xu, Jipeng Zhang, Renjie Pi, Xiaoyu Wang, Tong Zhang• 2024

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningGSM8K
Accuracy87.1
983
Mathematical ReasoningGSM8K (test)
Accuracy74.1
751
Mathematical ReasoningMATH
Accuracy59.8
643
Mathematical ReasoningMATH (test)
Overall Accuracy41.7
433
Instruction FollowingMT-Bench GPT-4o judge (test)
MT-Bench Score7.76
12
Mathematical ReasoningAIME 2024 (reference set)
Pass@133.3
4
Mathematical ReasoningAIME 2025 (reference set)
pass@126.7
4
Mathematical ReasoningAIMO 2025 (reference set)
Pass@120
4
Showing 8 of 8 rows

Other info

Code

Follow for update