Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

"Why Should I Trust You?": Explaining the Predictions of Any Classifier

About

Despite widespread adoption, machine learning models remain mostly black boxes. Understanding the reasons behind predictions is, however, quite important in assessing trust, which is fundamental if one plans to take action based on a prediction, or when choosing whether to deploy a new model. Such understanding also provides insights into the model, which can be used to transform an untrustworthy model or prediction into a trustworthy one. In this work, we propose LIME, a novel explanation technique that explains the predictions of any classifier in an interpretable and faithful manner, by learning an interpretable model locally around the prediction. We also propose a method to explain models by presenting representative individual predictions and their explanations in a non-redundant way, framing the task as a submodular optimization problem. We demonstrate the flexibility of these methods by explaining different models for text (e.g. random forests) and image classification (e.g. neural networks). We show the utility of explanations via novel experiments, both simulated and with human subjects, on various scenarios that require trust: deciding if one should trust a prediction, choosing between models, improving an untrustworthy classifier, and identifying why a classifier should not be trusted.

Marco Tulio Ribeiro, Sameer Singh, Carlos Guestrin• 2016

Related benchmarks

TaskDatasetResultRank
Intent ClassificationBanking77 (test)
Accuracy92.82
151
ExplainabilityImageNet (val)
Insertion69.4
104
Interpretation Error EvaluationImageNet
Interpretation Error19.95
80
InterpretationSST-2
L2 Norm0.0981
56
Interpretation errorIMDB (test)
L2 Norm Error0.0624
56
Local explanation fidelityWine
RMSE0.009
54
Local explanation fidelityBreast cancer n = 32
RMSE0.011
54
Local explanation fidelityDiabetes n = 9
RMSE0.018
54
Attribution FidelityImageNet 1,000 images (val)
µFidelity0.11
48
Interpretability EvaluationMS-COCO
Interpretation Error Rate22.64
40
Showing 10 of 99 rows
...

Other info

Code

Follow for update