Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

CRAFT: Concept Recursive Activation FacTorization for Explainability

About

Attribution methods, which employ heatmaps to identify the most influential regions of an image that impact model decisions, have gained widespread popularity as a type of explainability method. However, recent research has exposed the limited practical value of these methods, attributed in part to their narrow focus on the most prominent regions of an image -- revealing "where" the model looks, but failing to elucidate "what" the model sees in those areas. In this work, we try to fill in this gap with CRAFT -- a novel approach to identify both "what" and "where" by generating concept-based explanations. We introduce 3 new ingredients to the automatic concept extraction literature: (i) a recursive strategy to detect and decompose concepts across layers, (ii) a novel method for a more faithful estimation of concept importance using Sobol indices, and (iii) the use of implicit differentiation to unlock Concept Attribution Maps. We conduct both human and computer vision experiments to demonstrate the benefits of the proposed approach. We show that the proposed concept importance estimation technique is more faithful to the model than previous methods. When evaluating the usefulness of the method for human experimenters on a human-centered utility benchmark, we find that our approach significantly improves on two of the three test scenarios. Our code is freely available at github.com/deel-ai/Craft.

Thomas Fel, Agustin Picard, Louis Bethune, Thibaut Boissin, David Vigouroux, Julien Colin, R\'emi Cad\`ene, Thomas Serre• 2022

Related benchmarks

TaskDatasetResultRank
Explainability EvaluationHusky vs. Wolf
Session 1 Utility Score90.6
11
Explainability EvaluationLeaves
Session 186.2
11
Explainability EvaluationKit Fox vs. Red Fox
Session 1 Utility Score0.565
11
Object ClassificationCaltech-101 (test)
SURFMAE3.19
7
Object ClassificationFood-101
SURF MAE2.5
6
Multi-attribute predictionCelebA (test)
SURFMAE6.87
6
Object ClassificationImageNet InceptionV3 (test)
SURF MAE4.26
6
Predicting Model OutputOtter vs. Beaver
Accuracy (Session 1)86.3
5
Predicting Model OutputHusky vs. Wolf
Session 1 Accuracy55.5
5
Predicting Model OutputKit Fox vs. Red Fox
Accuracy (Session 1)76.8
5
Showing 10 of 10 rows

Other info

Code

Follow for update