Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Hierarchical Chain-of-Thought Prompting: Enhancing LLM Reasoning Performance and Efficiency

About

Chain-of-Thought (CoT) prompting has significantly improved the reasoning capabilities of large language models (LLMs). However, conventional CoT often relies on unstructured, flat reasoning chains that suffer from redundancy and suboptimal performance. In this work, we introduce Hierarchical Chain-of-Thought (Hi-CoT) prompting, a structured reasoning paradigm specifically designed to address the challenges of complex, multi-step reasoning. Hi-CoT decomposes the reasoning process into hierarchical substeps by alternating between instructional planning and step-by-step execution. This decomposition enables LLMs to better manage long reasoning horizons and maintain logical coherence. Extensive evaluations across diverse LLMs and mathematical reasoning benchmarks show that Hi-CoT consistently improves average accuracy by 6.2% (up to 61.4% on certain models and tasks) while reducing reasoning trace length by 13.9% compared to CoT prompting. We further show that accuracy and efficiency are maximized when models strictly adhere to the hierarchical structure. Our code is available at https://github.com/XingshuaiHuang/Hi-CoT.

Xingshuai Huang, Derek Li, Bahareh Nikpour, Parsa Omidi• 2026

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningMinerva Math
Accuracy39.7
209
Mathematical ReasoningAMC
Accuracy (%)66.3
134
Mathematical ReasoningMATH 500
Top-1 Accuracy90
112
Showing 3 of 3 rows

Other info

Follow for update