Trade-off Between Efficiency and Consistency for Removal-based Explanations
About
In the current landscape of explanation methodologies, most predominant approaches, such as SHAP and LIME, employ removal-based techniques to evaluate the impact of individual features by simulating various scenarios with specific features omitted. Nonetheless, these methods primarily emphasize efficiency in the original context, often resulting in general inconsistencies. In this paper, we demonstrate that such inconsistency is an inherent aspect of these approaches by establishing the Impossible Trinity Theorem, which posits that interpretability, efficiency, and consistency cannot hold simultaneously. Recognizing that the attainment of an ideal explanation remains elusive, we propose the utilization of interpretation error as a metric to gauge inefficiencies and inconsistencies. To this end, we present two novel algorithms founded on the standard polynomial basis, aimed at minimizing interpretation error. Our empirical findings indicate that the proposed methods achieve a substantial reduction in interpretation error, up to 31.8 times lower when compared to alternative techniques. Code is available at https://github.com/trusty-ai/efficient-consistent-explanations.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Interpretation Error Evaluation | ImageNet | Interpretation Error7.05 | 80 | |
| Interpretation | SST-2 | L2 Norm0.0434 | 56 | |
| Interpretation error | IMDB (test) | L2 Norm Error0.0175 | 56 | |
| Interpretability Evaluation | MS-COCO | Interpretation Error Rate1.83 | 40 | |
| Interpretation error | ImageNet (test) | L2 Norm0.1048 | 40 | |
| Interpretation error | ImageNet (val) | Interpretation Error0.1231 | 40 | |
| Model Interpretation | IMDB | C3 Truthful Gap0.174 | 8 | |
| Model Interpretation | ImageNet 1000 random images | C3 Truthful Gap0.246 | 8 |