Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Look at the Variance! Efficient Black-box Explanations with Sobol-based Sensitivity Analysis

About

We describe a novel attribution method which is grounded in Sensitivity Analysis and uses Sobol indices. Beyond modeling the individual contributions of image regions, Sobol indices provide an efficient way to capture higher-order interactions between image regions and their contributions to a neural network's prediction through the lens of variance. We describe an approach that makes the computation of these indices efficient for high-dimensional problems by using perturbation masks coupled with efficient estimators to handle the high dimensionality of images. Importantly, we show that the proposed method leads to favorable scores on standard benchmarks for vision (and language models) while drastically reducing the computing time compared to other black-box methods -- even surpassing the accuracy of state-of-the-art white-box methods which require access to internal representations. Our code is freely available: https://github.com/fel-thomas/Sobol-Attribution-Method

Thomas Fel, Remi Cadene, Mathieu Chalvidal, Matthieu Cord, David Vigouroux, Thomas Serre• 2021

Related benchmarks

TaskDatasetResultRank
ExplainabilityImageNet (val)
Insertion37
104
Attribution FidelityImageNet 1,000 images (val)
µFidelity0.23
48
DeletionImageNet 2,000 images (val)
Deletion Score0.147
48
Pointing localizationVOC 2007 (test)
Mean Accuracy (All)89.8
44
Pointing gameMSCOCO 2014 (val)
Mean Accuracy (All)57.5
42
Explanation FaithfulnessIMDB Review 1,000 sentences (val)
Word Deletion Score66.2
14
Showing 6 of 6 rows

Other info

Code

Follow for update