Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

What to Hide from Your Students: Attention-Guided Masked Image Modeling

About

Transformers and masked language modeling are quickly being adopted and explored in computer vision as vision transformers and masked image modeling (MIM). In this work, we argue that image token masking differs from token masking in text, due to the amount and correlation of tokens in an image. In particular, to generate a challenging pretext task for MIM, we advocate a shift from random masking to informed masking. We develop and exhibit this idea in the context of distillation-based MIM, where a teacher transformer encoder generates an attention map, which we use to guide masking for the student. We thus introduce a novel masking strategy, called attention-guided masking (AttMask), and we demonstrate its effectiveness over random masking for dense distillation-based MIM as well as plain distillation-based self-supervised learning on classification tokens. We confirm that AttMask accelerates the learning process and improves the performance on a variety of downstream tasks. We provide the implementation code at https://github.com/gkakogeorgiou/attmask.

Ioannis Kakogeorgiou, Spyros Gidaris, Bill Psomas, Yannis Avrithis, Andrei Bursuc, Konstantinos Karantzalos, Nikos Komodakis• 2022

Related benchmarks

TaskDatasetResultRank
Object DetectionCOCO 2017 (val)
AP25.9
2643
Instance SegmentationCOCO 2017 (val)--
1201
Video Object SegmentationDAVIS 2017 (val)
J mean52.5
1193
Semantic segmentationADE20K
mIoU45.3
1024
Image ClassificationImageNet-1K--
600
Instance SegmentationCOCO
APmask42
291
Object DetectionCOCO--
237
Video Object SegmentationDAVIS 2017
Jaccard Index (J)52.5
82
Text-to-motion generationHumanML3D
R-Precision (Top 1)49.9
64
Image ClassificationImageNet-1k (val)
Accuracy35.9
59
Showing 10 of 16 rows

Other info

Follow for update