Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Generic Attention-model Explainability for Interpreting Bi-Modal and Encoder-Decoder Transformers

About

Transformers are increasingly dominating multi-modal reasoning tasks, such as visual question answering, achieving state-of-the-art results thanks to their ability to contextualize information using the self-attention and co-attention mechanisms. These attention modules also play a role in other computer vision tasks including object detection and image segmentation. Unlike Transformers that only use self-attention, Transformers with co-attention require to consider multiple attention maps in parallel in order to highlight the information that is relevant to the prediction in the model's input. In this work, we propose the first method to explain prediction by any Transformer-based architecture, including bi-modal Transformers and Transformers with co-attentions. We provide generic solutions and apply these to the three most commonly used of these architectures: (i) pure self-attention, (ii) self-attention combined with co-attention, and (iii) encoder-decoder attention. We show that our method is superior to all existing methods which are adapted from single modality explainability.

Hila Chefer, Shir Gur, Lior Wolf• 2021

Related benchmarks

TaskDatasetResultRank
Image-to-Text RetrievalMS-COCO (test)
R@120.97
99
Anomaly SegmentationMVTec-AD (test)--
85
LocalizationImageNet-1k (val)
EHR0.297
79
Feature Importance AssessmentImageNet-1k (val)
Insertion Score34.71
78
Text-to-Image RetrievalMS-COCO (test)
R@115.37
66
Weakly Supervised Object LocalizationCUB-200-2011 (test)
Accuracy68.01
38
Feature Attribution EvaluationImageNet-1k (val)
MoRF Score30.06
33
Phrase LocalizationVisualGenome (VG) (test)
Pointing Accuracy54.72
29
SegmentationImageNet segmentation
Pixel Accuracy78.17
22
Explanation FaithfulnessImageNet 2015 (test)
AOPC0.707
22
Showing 10 of 51 rows

Other info

Follow for update