Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Nexus: Same Pretraining Loss, Better Downstream Generalization via Common Minima

About

Pretraining is the cornerstone of Large Language Models (LLMs), dominating the vast majority of computational budget and data to serve as the primary engine for their capabilities. During pretraining, LLMs acquire foundational knowledge from an unprecedentedly massive and diverse data sources, encompassing a vast array of domains such as general language, mathematics, code, and complex reasoning. In this work, we investigate an interesting geometric question regarding the converged state of pretraining: Does the model converge to a common minimizer across all data sources (e.g., \cref{fig:cwa_illustration:close}), or merely a minimizer of the summed loss (e.g., \cref{fig:cwa_illustration:distant})? We hypothesize that the geometric "closeness" of task-specific minima is intrinsically linked to downstream generalization. We reveal that standard optimizers (e.g., AdamW) often converge to points where task-specific minima are distant from each other. To address this, we propose the Nexus optimizer, which encourages the closeness of these minima by maximizing gradient similarity during optimization. Experiments across models ranging from 130M to 3B parameters, various data mixtures and hyperparameter schedules, show that Nexus \textit{significantly boosts downstream performance}, despite \textit{achieving the same pretraining loss} (see \cref{fig:demo:benchmark}). Notably, on the 3B model, Nexus reduces the out-of-distribution loss by 0.012 and yields up to a 15.0\% accuracy improvement on complex reasoning tasks (e.g., GSM8k). This finding challenges the reliance on pretraining loss as the sole proxy for model evaluation and demonstrates the importance of implicit biases in unlocking downstream generalization.

Huanran Chen, Huaqing Zhang, Xiao Li, Yinpeng Dong, Ke Shen, Jun Zhu• 2026

Related benchmarks

TaskDatasetResultRank
Multiple-choice Question AnsweringMMLU
Accuracy48.9
185
Language ModelingPre-training (val)--
13
Mathematical ReasoningGSM8K
Accuracy59
11
Language ModelingPre-training corpus
Loss1.602
9
Language ModelingOOD
Loss1.29
7
Aggregated PerformanceDownstream Average All
Accuracy40.3
4
Code GenerationHumanEval
Accuracy63
4
Graduate-level Science Question AnsweringGPQA D
Accuracy23.4
4
Language ModelingPublic Pretraining Dataset OOD
Loss1.606
4
Mathematics Problem SolvingMATH
Accuracy40
4
Showing 10 of 25 rows

Other info

Follow for update