STEEX: Steering Counterfactual Explanations with Semantics
About
As deep learning models are increasingly used in safety-critical applications, explainability and trustworthiness become major concerns. For simple images, such as low-resolution face portraits, synthesizing visual counterfactual explanations has recently been proposed as a way to uncover the decision mechanisms of a trained classification model. In this work, we address the problem of producing counterfactual explanations for high-quality images and complex scenes. Leveraging recent semantic-to-image models, we propose a new generative counterfactual explanation framework that produces plausible and sparse modifications which preserve the overall scene structure. Furthermore, we introduce the concept of "region-targeted counterfactual explanations", and a corresponding framework, where users can guide the generation of counterfactuals by specifying a set of semantic regions of the query image the explanation must be about. Extensive experiments are conducted on challenging datasets including high-quality portraits (CelebAMask-HQ) and driving scenes (BDD100k). Code is available at https://github.com/valeoai/STEEX
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Visual Counterfactual Explanation (Age) | CelebA Standard | FID11.8 | 11 | |
| Visual Counterfactual Explanation (Smile) | CelebA Standard | FID10.2 | 11 | |
| Counterfactual Visual Explanation | BDD100K | FID58.8 | 10 | |
| Visual Counterfactual Explanation (Age) | CelebA-HQ | FID26.8 | 9 | |
| Visual Counterfactual Explanation (Smile) | CelebA-HQ | FID21.9 | 9 | |
| Counterfactual Visual Explanation (Age attribute) | CelebA (test) | FID11.8 | 6 | |
| Counterfactual Visual Explanation (Smile attribute) | CelebA (test) | FID10.2 | 6 | |
| Counterfactual Explanation (Age) | CelebA-HQ (test) | FID26.8 | 5 | |
| Counterfactual Explanation (Smile) | CelebA-HQ (test) | FID21.9 | 5 |