Focused Chain-of-Thought: Efficient LLM Reasoning via Structured Input Information
About
Recent large language models achieve strong reasoning performance by generating detailed chain-of-thought traces, but this often leads to excessive token use and high inference latency. Existing efficiency approaches typically focus on model-centric interventions, such as reinforcement learning or supervised fine-tuning, to reduce verbosity. In contrast, we propose a training-free, input-centric approach. Inspired by cognitive psychology, we introduce Focused Chain-of-Thought (F-CoT), which separates information extraction from the reasoning process. F-CoT first organizes the essential information from a query into a concise, structured context and then guides the model to reason exclusively over this context. By preventing attention to irrelevant details, F-CoT naturally produces shorter reasoning paths. On arithmetic word problems, F-CoT reduces generated tokens by 2-3x while maintaining accuracy comparable to standard zero-shot CoT. These results highlight structured input as a simple yet effective lever for more efficient LLM reasoning.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Mathematical Reasoning | GSM-Hard | Solve Rate77.03 | 162 | |
| Mathematical Reasoning | MATH 500 | Pass@5 Rate99 | 16 | |
| Mathematical Reasoning | AIME 2024 | Pass@583.33 | 16 | |
| Mathematical Reasoning | AIME 2025 | Pass@576.67 | 16 | |
| Mathematical Reasoning | SVAMP | Pass@595.67 | 16 |