Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

WeakTr: Exploring Plain Vision Transformer for Weakly-supervised Semantic Segmentation

About

This paper explores the properties of the plain Vision Transformer (ViT) for Weakly-supervised Semantic Segmentation (WSSS). The class activation map (CAM) is of critical importance for understanding a classification network and launching WSSS. We observe that different attention heads of ViT focus on different image areas. Thus a novel weight-based method is proposed to end-to-end estimate the importance of attention heads, while the self-attention maps are adaptively fused for high-quality CAM results that tend to have more complete objects. Besides, we propose a ViT-based gradient clipping decoder for online retraining with the CAM results to complete the WSSS task. We name this plain Transformer-based Weakly-supervised learning framework WeakTr. It achieves the state-of-the-art WSSS performance on standard benchmarks, i.e., 78.4% mIoU on the val set of PASCAL VOC 2012 and 50.3% mIoU on the val set of COCO 2014. Code is available at https://github.com/hustvl/WeakTr.

Lianghui Zhu, Yingyue Li, Jiemin Fang, Yan Liu, Hao Xin, Wenyu Liu, Xinggang Wang• 2023

Related benchmarks

TaskDatasetResultRank
Semantic segmentationPASCAL VOC (val)
mIoU81.4
338
Semantic segmentationCOCO 2014 (val)
mIoU44.4
251
Semantic segmentationPascal VOC (test)
mIoU78.4
236
Semantic segmentationCOCO (val)
mIoU53.7
135
Showing 4 of 4 rows

Other info

Code

Follow for update