Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Reward-Guided Speculative Decoding for Efficient LLM Reasoning

About

We introduce Reward-Guided Speculative Decoding (RSD), a novel framework aimed at improving the efficiency of inference in large language models (LLMs). RSD synergistically combines a lightweight draft model with a more powerful target model, incorporating a controlled bias to prioritize high-reward outputs, in contrast to existing speculative decoding methods that enforce strict unbiasedness. RSD employs a process reward model to evaluate intermediate decoding steps and dynamically decide whether to invoke the target model, optimizing the trade-off between computational cost and output quality. We theoretically demonstrate that a threshold-based mixture strategy achieves an optimal balance between resource utilization and performance. Extensive evaluations on challenging reasoning benchmarks, including Olympiad-level tasks, show that RSD delivers significant efficiency gains against decoding with the target model only (up to 4.4x fewer FLOPs), while achieving significant better accuracy than parallel decoding method on average (up to +3.5). These results highlight RSD as a robust and cost-effective approach for deploying LLMs in resource-intensive scenarios. The code is available at https://github.com/BaohaoLiao/RSD.

Baohao Liao, Yuhui Xu, Hanze Dong, Junnan Li, Christof Monz, Silvio Savarese, Doyen Sahoo, Caiming Xiong• 2025

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningAMC 23
Accuracy67.5
198
Mathematical ReasoningMinerva--
138
Mathematical ReasoningOlympiad
Accuracy46.5
92
Mathematical ReasoningAIME 24
AIME 24 Accuracy16.67
84
Mathematical ReasoningMATH500
Accuracy79
24
Mathematical ReasoningOlympiadBench
Accuracy0.335
24
Mathematical ReasoningAMC23
Accuracy50
24
Knowledge ReasoningGPQA
Accuracy50.5
11
Showing 8 of 8 rows

Other info

Follow for update