Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Structure Enables Effective Self-Localization of Errors in LLMs

About

Self-correction in language models remains elusive. In this work, we explore whether language models can explicitly localize errors in incorrect reasoning, as a path toward building AI systems that can effectively correct themselves. We introduce a prompting method that structures reasoning as discrete, semantically coherent thought steps, and show that models are able to reliably localize errors within this structure, while failing to do so in conventional, unstructured chain-of-thought reasoning. Motivated by how the human brain monitors errors at discrete decision points and resamples alternatives, we introduce Iterative Correction Sampling of Thoughts (Thought-ICS), a self-correction framework. Thought-ICS iteratively prompts the model to generate reasoning one discrete and complete thought at a time--where each thought represents a deliberate decision by the model--creating natural boundaries for precise error localization. Upon verification, the model localizes the first erroneous step, and the system backtracks to generate alternative reasoning from the last correct point. When asked to correct reasoning verified as incorrect by an oracle, Thought-ICS achieves 20-40% self-correction lift. In a completely autonomous setting without external verification, it outperforms contemporary self-correction baselines.

Ankur Samanta, Akshayaa Magesh, Ayush Jain, Kavosh Asadi, Youliang Yu, Daniel Jiang, Boris Vidolov, Kaveh Hassani, Paul Sajda, Jalaj Bhandari, Yonathan Efroni• 2026

Related benchmarks

TaskDatasetResultRank
Commonsense ReasoningCSQA
Accuracy96
366
Mathematical ReasoningAIME
AIME Accuracy72
283
Question AnsweringGPQA
Accuracy69
258
Science ReasoningGPQA
Accuracy79
218
Mathematical ReasoningAMC 23
Accuracy92.5
198
Mathematical ReasoningMathQA
Accuracy90
95
Mathematical ReasoningMATH L5
Accuracy0.75
86
Showing 7 of 7 rows

Other info

Follow for update