Modeling Causal Mechanisms with Diffusion Models for Interventional and Counterfactual Queries
About
We consider the problem of answering observational, interventional, and counterfactual queries in a causally sufficient setting where only observational data and the causal graph are available. Utilizing the recent developments in diffusion models, we introduce diffusion-based causal models (DCM) to learn causal mechanisms, that generate unique latent encodings. These encodings enable us to directly sample under interventions and perform abduction for counterfactuals. Diffusion models are a natural fit here, since they can encode each node to a latent representation that acts as a proxy for exogenous noise. Our empirical evaluations demonstrate significant improvements over existing state-of-the-art methods for answering causal queries. Furthermore, we provide theoretical results that offer a methodology for analyzing counterfactual estimation in general encoder-decoder models, which could be useful in settings beyond our proposed approach.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Causal Generative Modeling | 11 synthetic datasets Mixed type variables (test) | RF Observed Accuracy0.58 | 7 | |
| Constrained Generation | Experimental Grid Oracle SEM (All experiments) | Score91.6 | 7 | |
| Causal Generative Modeling | 11 synthetic datasets Continuous variables (test) | RF Accuracy (Observed)56 | 5 | |
| Causal Inference | Sachs Additive Noise Semi-synthetic | RF Accuracy (Observed)56 | 5 | |
| Counterfactual reasoning | Chain NADD | MSE865 | 5 | |
| Counterfactual reasoning | Triangle NADD | MSE7.05e+3 | 5 | |
| Counterfactual reasoning | Diamond NADD | MSE1.36e+3 | 5 | |
| Causal Inference | Sachs Non-additive Noise Semi-synthetic | Accuracy (RF, Observed)59 | 5 | |
| Counterfactual reasoning | Y-struct NADD | MSE1.93e+4 | 5 | |
| Observational query | HDM-APE SO2 (test) | MMD^21.13 | 3 |