Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Learning to Track Instances without Video Annotations

About

Tracking segmentation masks of multiple instances has been intensively studied, but still faces two fundamental challenges: 1) the requirement of large-scale, frame-wise annotation, and 2) the complexity of two-stage approaches. To resolve these challenges, we introduce a novel semi-supervised framework by learning instance tracking networks with only a labeled image dataset and unlabeled video sequences. With an instance contrastive objective, we learn an embedding to discriminate each instance from the others. We show that even when only trained with images, the learned feature representation is robust to instance appearance variations, and is thus able to track objects steadily across frames. We further enhance the tracking capability of the embedding by learning correspondence from unlabeled videos in a self-supervised manner. In addition, we integrate this module into single-stage instance segmentation and pose estimation frameworks, which significantly reduce the computational complexity of tracking compared to two-stage networks. We conduct experiments on the YouTube-VIS and PoseTrack datasets. Without any video annotation efforts, our proposed method can achieve comparable or even better performance than most fully-supervised methods.

Yang Fu, Sifei Liu, Umar Iqbal, Shalini De Mello, Humphrey Shi, Jan Kautz• 2021

Related benchmarks

TaskDatasetResultRank
Video Instance SegmentationYouTube-VIS 2019 (val)
AP38.3
567
Video Instance SegmentationYTVIS 2019 (test val)
AP30.6
28
Multi-person pose trackingPoseTrack 2018 (val)
mAP71.4
25
Pose TrackingPoseTrack 2018 (val)
FPS4.1
4
Showing 4 of 4 rows

Other info

Code

Follow for update