Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

On the Eligibility of LLMs for Counterfactual Reasoning: A Decompositional Study

About

Counterfactual reasoning has emerged as a crucial technique for generalizing the reasoning capabilities of large language models (LLMs). By generating and analyzing counterfactual scenarios, researchers can assess the adaptability and reliability of model decision-making. Although prior work has shown that LLMs often struggle with counterfactual reasoning, it remains unclear which factors most significantly impede their performance across different tasks and modalities. In this paper, we propose a decompositional strategy that breaks down the counterfactual generation from causality construction to the reasoning over counterfactual interventions. To support decompositional analysis, we investigate \ntask datasets spanning diverse tasks, including natural language understanding, mathematics, programming, and vision-language tasks. Through extensive evaluations, we characterize LLM behavior across each decompositional stage and identify how modality type and intermediate reasoning influence performance. By establishing a structured framework for analyzing counterfactual reasoning, this work contributes to the development of more reliable LLM-based reasoning systems and informs future elicitation strategies.

Shuai Yang, Qi Yang, Luoxi Tang, Yuqiao Meng, Nancy Guo, Jeremy Blackburn, Zhaohan Xi• 2025

Related benchmarks

TaskDatasetResultRank
Causal Variable IdentificationCRASS--
7
Causal Variable IdentificationCLOMO--
7
Causal Variable IdentificationRNN-Topo--
7
Causal Variable IdentificationCVQA-Bool--
7
Causal Variable IdentificationCVQA Count--
7
Causal Variable IdentificationCOCO--
7
Causal Variable IdentificationArithmetic--
7
Causal Variable IdentificationMalAlgoQA--
7
Causal Variable IdentificationHumanEval Exe--
7
Causal Variable IdentificationOpen-Critic--
7
Showing 10 of 22 rows

Other info

Follow for update