Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Learning Important Features Through Propagating Activation Differences

About

The purported "black box" nature of neural networks is a barrier to adoption in applications where interpretability is essential. Here we present DeepLIFT (Deep Learning Important FeaTures), a method for decomposing the output prediction of a neural network on a specific input by backpropagating the contributions of all neurons in the network to every feature of the input. DeepLIFT compares the activation of each neuron to its 'reference activation' and assigns contribution scores according to the difference. By optionally giving separate consideration to positive and negative contributions, DeepLIFT can also reveal dependencies which are missed by other approaches. Scores can be computed efficiently in a single backward pass. We apply DeepLIFT to models trained on MNIST and simulated genomic data, and show significant advantages over gradient-based methods. Video tutorial: http://goo.gl/qKb7pL, ICML slides: bit.ly/deeplifticmlslides, ICML talk: https://vimeo.com/238275076, code: http://goo.gl/RM8jvH.

Avanti Shrikumar, Peyton Greenside, Anshul Kundaje• 2017

Related benchmarks

TaskDatasetResultRank
LocalizationImageNet-1k (val)--
79
Faithfulness EvaluationTellMeWhy
AUC π-Soft-NS0.313
67
Faithfulness EvaluationWikiBio
AUC π-Soft-NS0.348
67
Feature Attribution PlausibilityMDACE (test)
P32.4
65
Feature Relevance EvaluationImageNet (test)
R (Feature Relevance)0.48
60
Faithfulness EvaluationMDACE (test)
Comp Score77
40
Feature Interaction AttributionDyck-2 15,000 corpus size (test)
Average Relative Ranks (ARR)0.542
34
Explanation PlausibilityMDACE bigger (test)
Precision33.9
32
XAI EvaluationImageNet-S (val)
Selection Score4.447
28
Faithfulness EvaluationBoolQ
AUC π-Soft-NS32.6
27
Showing 10 of 91 rows
...

Other info

Follow for update