Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

RelayGen: Intra-Generation Model Switching for Efficient Reasoning

About

Large reasoning models (LRMs) achieve strong performance on complex reasoning tasks by generating long, multi-step reasoning trajectories, but inference-time scaling incurs substantial deployment cost. A key challenge is that generation difficulty varies within a single output, whereas existing efficiency-oriented approaches either ignore this intra-generation variation or rely on supervised token-level routing with high system complexity. We present \textbf{RelayGen}, a training-free, segment-level runtime model switching framework that exploits difficulty variation in long-form reasoning. Through offline analysis of generation uncertainty using token probability margins, we show that coarse-grained segment-level control is sufficient to capture difficulty transitions within a reasoning trajectory. RelayGen identifies model-specific switch cues that signal transitions to lower-difficulty segments and dynamically delegates their continuation to a smaller model, while preserving high-difficulty reasoning on the large model. Across multiple reasoning benchmarks, RelayGen substantially reduces inference latency while preserving most of the accuracy of large models. When combined with speculative decoding, RelayGen achieves up to 2.2$\times$ end-to-end speedup with less than 2\% accuracy degradation, without requiring additional training or learned routing components.

Jiwon Song, Yoongon Kim, Jae-Joon Kim• 2026

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningMATH500 (test)--
381
Mathematical ReasoningAIME 2025 (test)
Pass@1 Rate68.33
47
Scientific ReasoningGPQA Diamond
Pass@163.64
32
Mathematical ReasoningAIME 2025 (random 5 problems)
Speedup2.2
5
Showing 4 of 4 rows

Other info

GitHub

Follow for update