Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Improving Chain-of-Thought Reasoning via Quasi-Symbolic Abstractions

About

Chain-of-Though (CoT) represents a common strategy for reasoning in Large Language Models (LLMs) by decomposing complex tasks into intermediate inference steps. However, explanations generated via CoT are susceptible to content biases that negatively affect their robustness and faithfulness. To mitigate existing limitations, recent work has proposed using logical formalisms coupled with external symbolic solvers. However, fully symbolic approaches possess the bottleneck of requiring a complete translation from natural language to formal languages, a process that affects efficiency and flexibility. To achieve a trade-off, this paper investigates methods to disentangle content from logical reasoning without a complete formalisation. In particular, we present QuaSAR (for Quasi-Symbolic Abstract Reasoning), a variation of CoT that guides LLMs to operate at a higher level of abstraction via quasi-symbolic explanations. Our framework leverages the capability of LLMs to formalise only relevant variables and predicates, enabling the coexistence of symbolic elements with natural language. We show the impact of QuaSAR for in-context learning and for constructing demonstrations to improve the reasoning capabilities of smaller models. Our experiments show that quasi-symbolic abstractions can improve CoT-based methods by up to 8% accuracy, enhancing robustness and consistency on challenging adversarial variations on both natural language (i.e. MMLU-Redux) and symbolic reasoning tasks (i.e., GSM-Symbolic).

Leonardo Ranaldi, Marco Valentino, Andr\`e Freitas• 2025

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningGSM8K
Accuracy96.5
983
Mathematical ReasoningMATH
Accuracy36.4
643
Mathematical ReasoningMATH (test)
Overall Accuracy79.5
433
Mathematical ReasoningSVAMP
Accuracy97
368
Graduate-level Question AnsweringGPQA
Accuracy55.4
114
Mathematical ReasoningMGSM
Accuracy66.9
114
Multilingual Mathematical ReasoningMGSM (test)
Accuracy93.4
57
Arithmetic ReasoningSVAMP
Accuracy (Overall)82.6
54
Question AnsweringMMLU-Redux
Accuracy90.2
42
Natural Language ReasoningDROP
Accuracy88.9
33
Showing 10 of 17 rows

Other info

Follow for update