Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Transformer Meets Tracker: Exploiting Temporal Context for Robust Visual Tracking

About

In video object tracking, there exist rich temporal contexts among successive frames, which have been largely overlooked in existing trackers. In this work, we bridge the individual video frames and explore the temporal contexts across them via a transformer architecture for robust object tracking. Different from classic usage of the transformer in natural language processing tasks, we separate its encoder and decoder into two parallel branches and carefully design them within the Siamese-like tracking pipelines. The transformer encoder promotes the target templates via attention-based feature reinforcement, which benefits the high-quality tracking model generation. The transformer decoder propagates the tracking cues from previous templates to the current frame, which facilitates the object searching process. Our transformer-assisted tracking framework is neat and trained in an end-to-end manner. With the proposed transformer, a simple Siamese matching approach is able to outperform the current top-performing trackers. By combining our transformer with the recent discriminative tracking pipeline, our method sets several new state-of-the-art records on prevalent tracking benchmarks.

Ning Wang, Wengang Zhou, Jie Wang, Houqaing Li• 2021

Related benchmarks

TaskDatasetResultRank
Visual Object TrackingTrackingNet (test)
Normalized Precision (Pnorm)83.5
460
Visual Object TrackingLaSOT (test)
AUC66.5
444
Visual Object TrackingGOT-10k (test)
Average Overlap68.8
378
Object TrackingLaSoT
AUC63.9
333
Object TrackingTrackingNet
Precision (P)73.1
225
Visual Object TrackingGOT-10k
AO68.8
223
Visual Object TrackingUAV123 (test)
AUC67.5
188
Visual Object TrackingUAV123
AUC0.675
165
Visual Object TrackingVOT 2020 (test)
EAO0.3
147
Visual Object TrackingOTB-100
AUC71.1
136
Showing 10 of 41 rows

Other info

Code

Follow for update