Counterfactual Explanations on Robust Perceptual Geodesics
About
Latent-space optimization methods for counterfactual explanations - framed as minimal semantic perturbations that change model predictions - inherit the ambiguity of Wachter et al.'s objective: the choice of distance metric dictates whether perturbations are meaningful or adversarial. Existing approaches adopt flat or misaligned geometries, leading to off-manifold artifacts, semantic drift, or adversarial collapse. We introduce Perceptual Counterfactual Geodesics (PCG), a method that constructs counterfactuals by tracing geodesics under a perceptually Riemannian metric induced from robust vision features. This geometry aligns with human perception and penalizes brittle directions, enabling smooth, on-manifold, semantically valid transitions. Experiments on three vision datasets show that PCG outperforms baselines and reveals failure modes hidden under standard metrics.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Counterfactual Generation | AFHQ | L1 Distance0.79 | 5 | |
| Counterfactual Generation | FFHQ | L1 Distance0.42 | 5 | |
| Counterfactual Generation | PlantVillage | L1 Loss0.36 | 5 | |
| Counterfactual Generation | AFHQ STYLEGAN2 | FID8.3 | 5 |