Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Towards Efficient Large Language Reasoning Models via Extreme-Ratio Chain-of-Thought Compression

About

Chain-of-Thought (CoT) reasoning successfully enhances the reasoning capabilities of Large Language Models (LLMs), yet it incurs substantial computational overhead for inference. Existing CoT compression methods often suffer from a critical loss of logical fidelity at high compression ratios, resulting in significant performance degradation. To achieve high-fidelity, fast reasoning, we propose a novel EXTreme-RAtio Chain-of-Thought Compression framework, termed Extra-CoT, which aggressively reduces the token budget while preserving answer accuracy. To generate reliable, high-fidelity supervision, we first train a dedicated semantically-preserved compressor on mathematical CoT data with fine-grained annotations. An LLM is then fine-tuned on these compressed pairs via a mixed-ratio supervised fine-tuning (SFT), teaching it to follow a spectrum of compression budgets and providing a stable initialization for reinforcement learning (RL). We further propose Constrained and Hierarchical Ratio Policy Optimization (CHRPO) to explicitly incentivize question-solving ability under lower budgets by a hierarchical reward. Experiments on three mathematical reasoning benchmarks show the superiority of Extra-CoT. For example, on MATH-500 using Qwen3-1.7B, Extra-CoT achieves over 73\% token reduction with an accuracy improvement of 0.6\%, significantly outperforming state-of-the-art (SOTA) methods.

Yuntian Tang, Bohan Jia, Wenxuan Huang, Lianyue Zhang, Jiao Xie, Wenxi Li, Wei Li, Jie Hu, Xinghao Chen, Rongrong Ji, Shaohui Lin• 2026

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningSVAMP
Accuracy93
368
Mathematical ReasoningMultiArith
Accuracy99.4
116
Mathematical ReasoningGSM8K
Tokens210
17
Mathematical ReasoningMATH 500
Tokens Used452
17
Mathematical ReasoningAMC 2023
Tokens675
17
Mathematical ReasoningMetaMath 1k
Token Count212
14
General Knowledge (STEM)MMLU STEM
Token Count1.27e+3
14
Showing 7 of 7 rows

Other info

Follow for update