Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Token-Budget-Aware LLM Reasoning

About

Reasoning is critical for large language models (LLMs) to excel in a wide range of tasks. While methods like Chain-of-Thought (CoT) reasoning and enhance LLM performance by decomposing problems into intermediate steps, they also incur significant overhead in token usage, leading to increased costs. We find that the reasoning process of current LLMs is unnecessarily lengthy and it can be compressed by including a reasonable token budget in the prompt, but the choice of token budget plays a crucial role in the actual compression effectiveness. We then propose a token-budget-aware LLM reasoning framework that dynamically adjusts the number of reasoning tokens based on the reasoning complexity of each problem. Experiments show that our method effectively reduces token costs in CoT reasoning with only a slight performance reduction, offering a practical solution to balance efficiency and accuracy in LLM reasoning. Code: https://github.com/GeniusHTX/TALE

Tingxu Han, Zhenting Wang, Chunrong Fang, Shiyu Zhao, Shiqing Ma, Zhenyu Chen• 2024

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningMATH500 (test)
Accuracy91.8
514
Mathematical ReasoningGSM8K--
499
Mathematical ReasoningAMC
Accuracy (ACC)93.3
203
Mathematical ReasoningAIME 2024
Accuracy71.1
151
Mathematical ReasoningAIME 24
Accuracy71.1
113
Mathematical ReasoningGSM8K (test)
Accuracy78.57
79
Mathematical ReasoningMATH 500
Accuracy (%)94
54
Mathematical ReasoningGSM8K (test)
Accuracy93.6
33
Question AnsweringGPQA
Accuracy41.52
33
Mathematical ReasoningGSM8K
Tokens188.8
30
Showing 10 of 34 rows

Other info

Code

Follow for update