Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Making Slow Thinking Faster: Compressing LLM Chain-of-Thought via Step Entropy

About

Large Language Models (LLMs) using Chain-of-Thought (CoT) prompting excel at complex reasoning but generate verbose thought processes with considerable redundancy, leading to increased inference costs and reduced efficiency. We introduce a novel CoT compression framework based on step entropy, a metric that quantifies \emph{the informational contribution of individual reasoning steps} to identify redundancy. Through theoretical analysis and extensive empirical validation on mathematical reasoning benchmarks, we demonstrate that steps with low entropy are indeed highly redundant. Our experiments reveal that an astonishing 80\% of low-entropy intermediate steps can be pruned with minor degradation in the final answer accuracy across DeepSeek-R1-7B, 14B and Qwen3-8B. This finding sharply contrasts with random or high-entropy pruning, which severely impairs reasoning performance. Building on this, we propose a novel two-stage training strategy combining Supervised Fine-Tuning (SFT) and Group Relative Policy Optimization (GRPO) reinforcement learning. This approach enables LLMs to autonomously learn to generate compressed COTs during inference by strategically incorporating [SKIP] tokens. Our method significantly improves LLM inference efficiency while preserving accuracy, paving the way for more scalable LLM deployments and a better understanding of their internal reasoning. The code and data are released in https://github.com/staymylove/COT_Compresstion_via_Step_entropy.

Zeju Li, Jianyuan Zhong, Ziyang Zheng, Xiangyu Wen, Zhijian Xu, Yingying Cheng, Fan Zhang, Qiang Xu• 2025

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningGSM8K
Accuracy94.46
351
Multimodal ReasoningWeMath
Accuracy61.6
43
Mathematical ReasoningAIME 2025
Accuracy76.92
38
Multimodal ReasoningMMStar
Accuracy58.1
29
Multimodal ReasoningMathVista
Accuracy43.9
29
Mathematical ReasoningAIME 2024
ACC-620
26
Mathematical ReasoningGSM8K
Accuracy87.6
24
Mathematical ReasoningAIME 2025
Accuracy0.7692
24
Mathematical ReasoningMATH 500
Accuracy71.2
24
Mathematical ReasoningAMC 23
Accuracy48.5
24
Showing 10 of 24 rows

Other info

Follow for update