Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

MeMOTR: Long-Term Memory-Augmented Transformer for Multi-Object Tracking

About

As a video task, Multiple Object Tracking (MOT) is expected to capture temporal information of targets effectively. Unfortunately, most existing methods only explicitly exploit the object features between adjacent frames, while lacking the capacity to model long-term temporal information. In this paper, we propose MeMOTR, a long-term memory-augmented Transformer for multi-object tracking. Our method is able to make the same object's track embedding more stable and distinguishable by leveraging long-term memory injection with a customized memory-attention layer. This significantly improves the target association ability of our model. Experimental results on DanceTrack show that MeMOTR impressively surpasses the state-of-the-art method by 7.9% and 13.0% on HOTA and AssA metrics, respectively. Furthermore, our model also outperforms other Transformer-based methods on association performance on MOT17 and generalizes well on BDD100K. Code is available at https://github.com/MCG-NJU/MeMOTR.

Ruopeng Gao, Limin Wang• 2023

Related benchmarks

TaskDatasetResultRank
Multiple Object TrackingMOT17 (test)
MOTA72.8
921
Multi-Object TrackingDanceTrack (test)
HOTA0.685
355
Multi-Object TrackingSportsMOT (test)
HOTA70
199
Multi-Object TrackingBDD100K (val)--
70
Multi-Object TrackingMOT17
MOTA72.8
55
Multi-Object TrackingDanceTrack 58 (test)
HOTA63.4
20
Multi-Object TrackingSportsMOT 11 (test)
HOTA68.8
13
Showing 7 of 7 rows

Other info

Code

Follow for update