Exogenous Matching: Learning Good Proposals for Tractable Counterfactual Estimation
About
We propose an importance sampling method for tractable and efficient estimation of counterfactual expressions in general settings, named Exogenous Matching. By minimizing a common upper bound of counterfactual estimators, we transform the variance minimization problem into a conditional distribution learning problem, enabling its integration with existing conditional distribution modeling approaches. We validate the theoretical results through experiments under various types and settings of Structural Causal Models (SCMs) and demonstrate the outperformance on counterfactual estimation tasks compared to other existing importance sampling methods. We also explore the impact of injecting structural prior knowledge (counterfactual Markov boundaries) on the results. Finally, we apply this method to identifiable proxy SCMs and demonstrate the unbiasedness of the estimates, empirically illustrating the applicability of the method to practical scenarios.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Counterfactual Sampling | SIMPSON-NLIN Markovian diffeomorphic (test) | ESP2 | 15 | |
| Counterfactual Sampling | NAPKIN Semi-Markovian continuous (test) | ESP0.009 | 15 | |
| Counterfactual Sampling | FAIRNESS-XW Regional canonical (test) | ESP23.1 | 15 | |
| Counterfactual effect estimation | NCM FAIRNESS-XY | ATE0.00e+0 | 6 | |
| Counterfactual effect estimation | NCM FAIRNESS-XW (original proxy comparison) | ATE0.00e+0 | 6 | |
| Counterfactual effect estimation | NCM FAIRNESS-XY (original_proxy_comparison) | ATE0.00e+0 | 6 | |
| Counterfactual effect estimation | NCM FAIRNESS-YW (original_proxy_comparison) | ATE0.04 | 6 | |
| Counterfactual Density Estimation | SIMPSON-NLIN | FR (|s|=1)1 | 6 | |
| Counterfactual density kernel estimation | SIMPSON-NLIN |s|=1 | FR1 | 6 | |
| Counterfactual density kernel estimation | TRIANGLE-NLIN (|s|=1) | FR15 | 6 |