Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Good Counterfactuals and Where to Find Them: A Case-Based Technique for Generating Counterfactuals for Explainable AI (XAI)

About

Recently, a groundswell of research has identified the use of counterfactual explanations as a potentially significant solution to the Explainable AI (XAI) problem. It is argued that (a) technically, these counterfactual cases can be generated by permuting problem-features until a class change is found, (b) psychologically, they are much more causally informative than factual explanations, (c) legally, they are GDPR-compliant. However, there are issues around the finding of good counterfactuals using current techniques (e.g. sparsity and plausibility). We show that many commonly-used datasets appear to have few good counterfactuals for explanation purposes. So, we propose a new case based approach for generating counterfactuals using novel ideas about the counterfactual potential and explanatory coverage of a case-base. The new technique reuses patterns of good counterfactuals, present in a case-base, to generate analogous counterfactuals that can explain new problems and their solutions. Several experiments show how this technique can improve the counterfactual potential and explanatory coverage of case-bases that were previously found wanting.

Mark T. Keane, Barry Smyth• 2020

Related benchmarks

TaskDatasetResultRank
Counterfactual Explanation GenerationWine
Validity1
9
Counterfactual Explanationsmoons
Coverage100
6
Counterfactual ExplanationsLaw
Coverage100
6
Counterfactual ExplanationsAudit
Coverage1
6
Counterfactual ExplanationsHELOC
Coverage100
6
Counterfactual Explanation GenerationBlobs
Coverage100
5
Counterfactual Explanation GenerationDigits
Coverage100
5
Counterfactual ExplanationsAdult
Coverage100
4
Counterfactual ExplanationsCredit-g
Coverage100
4
Counterfactual ExplanationsCredit-A
Coverage100
4
Showing 10 of 10 rows

Other info

Follow for update