Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Prompting for Multi-Modal Tracking

About

Multi-modal tracking gains attention due to its ability to be more accurate and robust in complex scenarios compared to traditional RGB-based tracking. Its key lies in how to fuse multi-modal data and reduce the gap between modalities. However, multi-modal tracking still severely suffers from data deficiency, thus resulting in the insufficient learning of fusion modules. Instead of building such a fusion module, in this paper, we provide a new perspective on multi-modal tracking by attaching importance to the multi-modal visual prompts. We design a novel multi-modal prompt tracker (ProTrack), which can transfer the multi-modal inputs to a single modality by the prompt paradigm. By best employing the tracking ability of pre-trained RGB trackers learning at scale, our ProTrack can achieve high-performance multi-modal tracking by only altering the inputs, even without any extra training on multi-modal data. Extensive experiments on 5 benchmark datasets demonstrate the effectiveness of the proposed ProTrack.

Jinyu Yang, Zhe Li, Feng Zheng, Ale\v{s} Leonardis, Jingkuan Song• 2022

Related benchmarks

TaskDatasetResultRank
RGB-T TrackingLasHeR (test)
PR53.8
244
RGB-T TrackingRGBT234 (test)
Precision Rate79.5
189
RGB-D Object TrackingVOT-RGBD 2022 (public challenge)
EAO65.1
167
RGB-D Object TrackingDepthTrack (test)
Precision58.3
145
RGB-T TrackingRGBT234
Precision79.5
98
RGBT TrackingRGBT234
PR78.6
65
Object TrackingVisEvent (test)
PR63.2
63
RGBT TrackingLasHeR
PR50.9
55
RGBT TrackingRGBT 234
Precision Rate78.6
53
Visual Object TrackingDepthTrack
Precision0.583
41
Showing 10 of 26 rows

Other info

Follow for update