Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Let it Calm: Exploratory Annealed Decoding for Verifiable Reinforcement Learning

About

Reinforcement learning with verifiable rewards (RLVR) is a powerful paradigm for enhancing the reasoning capabilities of large language models (LLMs), yet its success hinges on effective exploration. An ideal exploration strategy must navigate two fundamental challenges: it must preserve sample quality while also ensuring training stability. While standard fixed-temperature sampling is simple, it struggles to balance these competing demands, as high temperatures degrade sample quality and low temperatures limit discovery. In this work, we propose a simpler and more effective strategy, Exploratory Annealed Decoding (EAD), grounded in the insight that exploration is most impactful on early tokens which define a sequence's semantic direction. EAD implements an intuitive **explore-at-the-beginning, exploit-at-the-end** strategy by annealing the sampling temperature from high to low during generation. This dynamic schedule encourages meaningful, high-level diversity at the start, then gradually lowers the temperature to preserve sample quality and keep the sampling distribution close to the target policy, which is essential for stable training. We demonstrate that EAD is a lightweight, plug-and-play method that significantly improves sample efficiency, consistently outperforming fixed-temperature sampling across various RLVR algorithms and model sizes. Our work suggests that aligning exploration with the natural dynamics of sequential generation offers a robust path to improving LLM reasoning.

Chenghao Yang, Lin Gui, Chenxiao Yang, Victor Veitch, Lizhu Zhang, Zhuokai Zhao• 2025

Related benchmarks

TaskDatasetResultRank
Code GenerationHumanEval (test)
Pass@176.22
444
Scientific ReasoningGPQA Diamond (test)
Accuracy38.19
32
Mathematical ReasoningOlympiadBench
Avg Success Rate @843.69
23
Mathematical ReasoningMinerva Math
Avg@823.9
14
Mathematical ReasoningAverage Reasoning Benchmarks
Avg@839.66
14
Mathematical ReasoningAMC 23
Avg@851.88
14
Mathematical ReasoningMATH 500
Avg@8 Score80
14
Mathematical ReasoningOmni-MATH
Average Score @824.32
14
Mathematical ReasoningAIME 24
Avg@814.17
14
General Knowledge ReasoningMMLU Pro (test)
Accuracy36.22
10
Showing 10 of 10 rows

Other info

Follow for update