Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Straight to Zero: Why Linearly Decaying the Learning Rate to Zero Works Best for LLMs

About

LLMs are commonly trained with a learning rate (LR) warmup, followed by cosine decay to 10% of the maximum (10x decay). In a large-scale empirical study, we show that under an optimal peak LR, a simple linear decay-to-zero (D2Z) schedule consistently outperforms other schedules when training at compute-optimal dataset sizes. D2Z is superior across a range of model sizes, batch sizes, datasets, and vocabularies. Benefits increase as dataset size increases. Leveraging a novel interpretation of AdamW as an exponential moving average of weight updates, we show how linear D2Z optimally balances the demands of early training (moving away from initial conditions) and late training (averaging over more updates in order to mitigate gradient noise). In experiments, a 610M-parameter model trained for 80 tokens-per-parameter (TPP) using D2Z achieves lower loss than when trained for 200 TPP using 10x decay, corresponding to an astonishing 60% compute savings. Models such as Llama2-7B, trained for 286 TPP with 10x decay, could likely have saved a majority of compute by training with D2Z.

Shane Bergsma, Nolan Dey, Gurpreet Gosal, Gavia Gray, Daria Soboleva, Joel Hestness• 2025

Related benchmarks

TaskDatasetResultRank
Multitask Language UnderstandingMMLU
Accuracy46.8
413
Instruction FollowingAlpacaEval
Win Rate76.3
227
Logical reasoningBBH
Accuracy31
201
Mathematical ReasoningGSM8K
Math Score43.5
197
Reading ComprehensionDROP
DROP Accuracy19.6
111
Multitask KnowledgeMMLU
Accuracy35
53
General IntelligenceAGI-Eval
AGI Eval Score33.6
24
Reading ComprehensionDROP
DROP Score30.5
16
General Language Modeling PerformanceAggregate AlpacaEval, TruthfulQA, GSM8K, DROP, AGI Eval, BBH, MMLU
Average Score41.4
16
TruthfulnessTruthfulQA
TruthfulQA39.4
8
Showing 10 of 11 rows

Other info

Follow for update