Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Scaling Laws for Optimal Data Mixtures

About

Large foundation models are typically trained on data from multiple domains, with the data mixture--the proportion of each domain used--playing a critical role in model performance. The standard approach to selecting this mixture relies on trial and error, which becomes impractical for large-scale pretraining. We propose a systematic method to determine the optimal data mixture for any target domain using scaling laws. Our approach accurately predicts the loss of a model of size $N$ trained with $D$ tokens and a specific domain weight vector $h$. We validate the universality of these scaling laws by demonstrating their predictive power in three distinct and large-scale settings: large language model (LLM), native multimodal model (NMM), and large vision models (LVM) pretraining. We further show that these scaling laws can extrapolate to new data mixtures and across scales: their parameters can be accurately estimated using a few small-scale training runs, and used to estimate the performance at larger scales and unseen domain weights. The scaling laws allow to derive the optimal domain weights for any target domain under a given training budget ($N$,$D$), providing a principled alternative to costly trial-and-error methods.

Mustafa Shukor, Louis Bethune, Dan Busbridge, David Grangier, Enrico Fini, Alaaeldin El-Nouby, Pierre Ablin• 2025

Related benchmarks

TaskDatasetResultRank
Code GenerationHumanEval--
1036
Language UnderstandingMMLU
Accuracy56.6
825
Science Question AnsweringARC Challenge
Accuracy54.1
342
General ReasoningBBH
BBH General Reasoning Accuracy45.3
98
MathematicsMATH
MATH Accuracy39.9
85
Chinese Language UnderstandingC-Eval
Accuracy62.8
56
Aggregated LLM EvaluationBalanced Objective Aggregate Suite
Weighted Average Score52.6
5
Large Language Model EvaluationMath Specialized Target (test)
Weighted Average Score49.4
4
Large Language Model EvaluationCode Specialized Target (test)
Weighted Average Score52
4
Large Language Model EvaluationKnowledge Specialized Target (test)
Weighted Average Score55.9
4
Showing 10 of 10 rows

Other info

Follow for update