Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps
About
This paper addresses the visualisation of image classification models, learnt using deep Convolutional Networks (ConvNets). We consider two visualisation techniques, based on computing the gradient of the class score with respect to the input image. The first one generates an image, which maximises the class score [Erhan et al., 2009], thus visualising the notion of the class, captured by a ConvNet. The second technique computes a class saliency map, specific to a given image and class. We show that such maps can be employed for weakly supervised object segmentation using classification ConvNets. Finally, we establish the connection between the gradient-based ConvNet visualisation methods and deconvolutional networks [Zeiler et al., 2013].
Karen Simonyan, Andrea Vedaldi, Andrew Zisserman• 2013
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | MNIST (test) | Accuracy91.35 | 882 | |
| Image Classification | SVHN (test) | Accuracy63.74 | 362 | |
| Explainability | ImageNet (val) | Insertion36.3 | 104 | |
| Feature Relevance Evaluation | ImageNet (test) | R (Feature Relevance)0.4 | 60 | |
| Attribution Fidelity | ImageNet 1,000 images (val) | µFidelity0.192 | 48 | |
| Deletion | ImageNet 2,000 images (val) | Deletion Score0.174 | 48 | |
| Object Localisation | ILSVRC (val) | Top-1 Error34.83 | 44 | |
| Pointing localization | VOC 2007 (test) | Mean Accuracy (All)76.3 | 44 | |
| Pointing game | MSCOCO 2014 (val) | Mean Accuracy (All)37.7 | 42 | |
| Feature Attribution Evaluation | ImageNet standard (val) | AUC65 | 39 |
Showing 10 of 52 rows