Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

IDGI: A Framework to Eliminate Explanation Noise from Integrated Gradients

About

Integrated Gradients (IG) as well as its variants are well-known techniques for interpreting the decisions of deep neural networks. While IG-based approaches attain state-of-the-art performance, they often integrate noise into their explanation saliency maps, which reduce their interpretability. To minimize the noise, we examine the source of the noise analytically and propose a new approach to reduce the explanation noise based on our analytical findings. We propose the Important Direction Gradient Integration (IDGI) framework, which can be easily incorporated into any IG-based method that uses the Reimann Integration for integrated gradient computation. Extensive experiments with three IG-based methods show that IDGI improves them drastically on numerous interpretability metrics.

Ruo Yang, Binghui Wang, Mustafa Bilgic• 2023

Related benchmarks

TaskDatasetResultRank
SegmentationImageNet segmentation
Pixel Accuracy76.42
22
Explanation FaithfulnessImageNet 2015 (test)
AOPC0.706
22
Showing 2 of 2 rows

Other info

Follow for update