Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

SoftCoT: Soft Chain-of-Thought for Efficient Reasoning with LLMs

About

Chain-of-Thought (CoT) reasoning enables Large Language Models (LLMs) to solve complex reasoning tasks by generating intermediate reasoning steps. However, most existing approaches focus on hard token decoding, which constrains reasoning within the discrete vocabulary space and may not always be optimal. While recent efforts explore continuous-space reasoning, they often require full-model fine-tuning and suffer from catastrophic forgetting, limiting their applicability to state-of-the-art LLMs that already perform well in zero-shot settings with a proper instruction. To address this challenge, we propose a novel approach for continuous-space reasoning that does not require modifying the LLM. Specifically, we employ a lightweight fixed assistant model to speculatively generate instance-specific soft thought tokens as the initial chain of thoughts, which are then mapped into the LLM's representation space via a trainable projection module. Experimental results on five reasoning benchmarks demonstrate that our method enhances LLM reasoning performance through supervised, parameter-efficient fine-tuning. Source code is available at https://github.com/xuyige/SoftCoT.

Yige Xu, Xu Guo, Zhiwei Zeng, Chunyan Miao• 2025

Related benchmarks

TaskDatasetResultRank
Code GenerationHumanEval (test)
Pass@171.83
444
Code GenerationMBPP (test)
Pass@156.04
276
Mathematical ReasoningSVAMP (test)
Accuracy40
233
Mathematical ReasoningAQUA
Accuracy80.63
132
Commonsense ReasoningStrategyQA
Accuracy71.18
125
Commonsense ReasoningStrategyQA (test)
Accuracy60.61
81
Arithmetic ReasoningMultiArith (test)
Accuracy74.4
67
Mathematical ReasoningASDiv Aug (test)
Accuracy88.9
25
Mathematical ReasoningGSM8K-NL (test)
Accuracy36.8
19
Mathematical ReasoningASDiv-Aug
Accuracy92.14
15
Showing 10 of 14 rows

Other info

Code

Follow for update