Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Comprehensive Attention Self-Distillation for Weakly-Supervised Object Detection

About

Weakly Supervised Object Detection (WSOD) has emerged as an effective tool to train object detectors using only the image-level category labels. However, without object-level labels, WSOD detectors are prone to detect bounding boxes on salient objects, clustered objects and discriminative object parts. Moreover, the image-level category labels do not enforce consistent object detection across different transformations of the same images. To address the above issues, we propose a Comprehensive Attention Self-Distillation (CASD) training approach for WSOD. To balance feature learning among all object instances, CASD computes the comprehensive attention aggregated from multiple transformations and feature layers of the same images. To enforce consistent spatial supervision on objects, CASD conducts self-distillation on the WSOD networks, such that the comprehensive attention is approximated simultaneously by multiple transformations and feature layers of the same images. CASD produces new state-of-the-art WSOD results on standard benchmarks such as PASCAL VOC 2007/2012 and MS-COCO.

Zeyi Huang, Yang Zou, Vijayakumar Bhagavatula, Dong Huang• 2020

Related benchmarks

TaskDatasetResultRank
Object DetectionMS-COCO 2017 (val)--
237
Object DetectionMS-COCO (val)
mAP0.128
138
Object LocalizationPASCAL VOC 2007 (trainval)
CorLoc70.4
118
Object DetectionVOC 2007 (test)
AP@5056.8
52
Object LocalizationPASCAL VOC 2012 (trainval)
CorLoc72.3
51
Object DetectionVOC 2012 (test)--
25
Weakly Supervised Object DetectionPASCAL VOC 2007 (test)
mAP@0.556.8
17
Object DetectionCOCO 2017 (test)
Overall AP13.9
10
Weakly Supervised Object DetectionPASCAL VOC 2012 (test)
mAP@0.553.6
8
Showing 9 of 9 rows

Other info

Code

Follow for update