Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Amortized Inference of Causal Models via Conditional Fixed-Point Iterations

About

Structural Causal Models (SCMs) offer a principled framework to reason about interventions and support out-of-distribution generalization, which are key goals in scientific discovery. However, the task of learning SCMs from observed data poses formidable challenges, and often requires training a separate model for each dataset. In this work, we propose an amortized inference framework that trains a single model to predict the causal mechanisms of SCMs conditioned on their observational data and causal graph. We first use a transformer-based architecture for amortized learning of dataset embeddings, and then extend the Fixed-Point Approach (FiP) to infer the causal mechanisms conditionally on their dataset embeddings. As a byproduct, our method can generate observational and interventional data from novel SCMs at inference time, without updating parameters. Empirical results show that our amortized procedure performs on par with baselines trained specifically for each dataset on both in and out-of-distribution problems, and also outperforms them in scarce data regimes.

Divyat Mahajan, Jannes Gladrow, Agrin Hilmkil, Cheng Zhang, Meyer Scetbon• 2024

Related benchmarks

TaskDatasetResultRank
Noise PredictionAVICI (out-of-distribution)
LIN (RMSE)0.07
32
Sample GenerationAVICI LIN (Out-of-distribution)
RMSE0.05
16
Sample GenerationAVICI RFF (Out-of-distribution)
RMSE0.08
16
Sample GenerationAVICI LIN (In-distribution)
RMSE0.05
16
Sample GenerationAVICI RFF (In-distribution)
RMSE0.14
16
Interventional GenerationAVICI In-distribution
LIN RMSE0.09
16
Noise PredictionAVICI In-distribution
LIN RMSE0.05
16
Counterfactual GenerationAVICI (test)
LIN RMSE (IN)0.09
16
Generating observational dataecoli
MMD (Generated vs Query)0.019
8
Noise PredictionAVICI n=400 (test)
LIN (IN)6
8
Showing 10 of 12 rows

Other info

Follow for update