Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Generating Interpretable Counterfactual Explanations By Implicit Minimisation of Epistemic and Aleatoric Uncertainties

About

Counterfactual explanations (CEs) are a practical tool for demonstrating why machine learning classifiers make particular decisions. For CEs to be useful, it is important that they are easy for users to interpret. Existing methods for generating interpretable CEs rely on auxiliary generative models, which may not be suitable for complex datasets, and incur engineering overhead. We introduce a simple and fast method for generating interpretable CEs in a white-box setting without an auxiliary model, by using the predictive uncertainty of the classifier. Our experiments show that our proposed algorithm generates more interpretable CEs, according to IM1 scores, than existing methods. Additionally, our approach allows us to estimate the uncertainty of a CE, which may be important in safety-critical applications, such as those in the medical domain.

Lisa Schut, Oscar Key, Rory McGrath, Luca Costabello, Bogdan Sacaleanu, Medb Corcoran, Yarin Gal• 2021

Related benchmarks

TaskDatasetResultRank
Counterfactual ExplanationsPneumoniaMNIST (test)
IM11.157
16
Counterfactual ExplanationsSpam (test)
IM11.2756
16
Counterfactual ExplanationsCredit (test)
IM10.9251
16
Counterfactual ExplanationsBreast Cancer (test)
IM11.1151
16
Counterfactual ExplanationsMNIST (test)
IM1 Score1.2373
11
Counterfactual ExplanationMNIST (test)
Validity Score96
10
Showing 6 of 6 rows

Other info

Follow for update