Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

AiATrack: Attention in Attention for Transformer Visual Tracking

About

Transformer trackers have achieved impressive advancements recently, where the attention mechanism plays an important role. However, the independent correlation computation in the attention mechanism could result in noisy and ambiguous attention weights, which inhibits further performance improvement. To address this issue, we propose an attention in attention (AiA) module, which enhances appropriate correlations and suppresses erroneous ones by seeking consensus among all correlation vectors. Our AiA module can be readily applied to both self-attention blocks and cross-attention blocks to facilitate feature aggregation and information propagation for visual tracking. Moreover, we propose a streamlined Transformer tracking framework, dubbed AiATrack, by introducing efficient feature reuse and target-background embeddings to make full use of temporal references. Experiments show that our tracker achieves state-of-the-art performance on six tracking benchmarks while running at a real-time speed.

Shenyuan Gao, Chunluan Zhou, Chao Ma, Xinggang Wang, Junsong Yuan• 2022

Related benchmarks

TaskDatasetResultRank
Visual Object TrackingTrackingNet (test)
Normalized Precision (Pnorm)87.8
460
Visual Object TrackingLaSOT (test)
AUC69
444
Visual Object TrackingGOT-10k (test)
Average Overlap69.6
378
Object TrackingLaSoT
AUC69.6
333
RGB-T TrackingLasHeR (test)
PR46.3
244
Object TrackingTrackingNet
Precision (P)80.4
225
Visual Object TrackingGOT-10k
AO69.6
223
Visual Object TrackingUAV123 (test)
AUC70.6
188
RGB-D Object TrackingVOT-RGBD 2022 (public challenge)
EAO0.641
167
Visual Object TrackingVOT 2020 (test)
EAO0.53
147
Showing 10 of 33 rows

Other info

Code

Follow for update