Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Fast Convergence of DETR with Spatially Modulated Co-Attention

About

The recently proposed Detection Transformer (DETR) model successfully applies Transformer to objects detection and achieves comparable performance with two-stage object detection frameworks, such as Faster-RCNN. However, DETR suffers from its slow convergence. Training DETR from scratch needs 500 epochs to achieve a high accuracy. To accelerate its convergence, we propose a simple yet effective scheme for improving the DETR framework, namely Spatially Modulated Co-Attention (SMCA) mechanism. The core idea of SMCA is to conduct location-aware co-attention in DETR by constraining co-attention responses to be high near initially estimated bounding box locations. Our proposed SMCA increases DETR's convergence speed by replacing the original co-attention mechanism in the decoder while keeping other operations in DETR unchanged. Furthermore, by integrating multi-head and scale-selection attention designs into SMCA, our fully-fledged SMCA can achieve better performance compared to DETR with a dilated convolution-based backbone (45.6 mAP at 108 epochs vs. 43.3 mAP at 500 epochs). We perform extensive ablation studies on COCO dataset to validate SMCA. Code is released at https://github.com/gaopengcuhk/SMCA-DETR .

Peng Gao, Minghang Zheng, Xiaogang Wang, Jifeng Dai, Hongsheng Li• 2021

Related benchmarks

TaskDatasetResultRank
Object DetectionCOCO 2017 (val)
AP46.3
2454
Video Object DetectionImageNet VID (val)
mAP (%)53.5
341
Object DetectionMS-COCO 2017 (val)--
237
Object DetectionCOCO (minival)
mAP44.4
184
Video Moment RetrievalCharades-STA (test)
Recall@1 (IoU=0.5)35.48
77
Video Moment RetrievalTACOS (test)
Recall@1 (0.5 Threshold)29.12
70
Showing 6 of 6 rows

Other info

Follow for update