Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

TransMOT: Spatial-Temporal Graph Transformer for Multiple Object Tracking

About

Tracking multiple objects in videos relies on modeling the spatial-temporal interactions of the objects. In this paper, we propose a solution named TransMOT, which leverages powerful graph transformers to efficiently model the spatial and temporal interactions among the objects. TransMOT effectively models the interactions of a large number of objects by arranging the trajectories of the tracked objects as a set of sparse weighted graphs, and constructing a spatial graph transformer encoder layer, a temporal transformer encoder layer, and a spatial graph transformer decoder layer based on the graphs. TransMOT is not only more computationally efficient than the traditional Transformer, but it also achieves better tracking accuracy. To further improve the tracking speed and accuracy, we propose a cascade association framework to handle low-score detections and long-term occlusions that require large computational resources to model in TransMOT. The proposed method is evaluated on multiple benchmark datasets including MOT15, MOT16, MOT17, and MOT20, and it achieves state-of-the-art performance on all the datasets.

Peng Chu, Jiang Wang, Quanzeng You, Haibin Ling, Zicheng Liu• 2021

Related benchmarks

TaskDatasetResultRank
Multiple Object TrackingMOT17 (test)
MOTA76.7
921
Multiple Object TrackingMOT20 (test)
MOTA77.5
358
Multi-Object TrackingMOT 2016 (test)
MOTA76.7
59
Multi-Object TrackingMOT 2020 (test)
MOTA77.5
44
Multi-Object TrackingMOT15 (test)
MOTA57
38
Multi-Object TrackingMOT20 Private detections (test)
IDF175.2
24
Showing 6 of 6 rows

Other info

Follow for update