Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Rethinking Self-supervised Correspondence Learning: A Video Frame-level Similarity Perspective

About

Learning a good representation for space-time correspondence is the key for various computer vision tasks, including tracking object bounding boxes and performing video object pixel segmentation. To learn generalizable representation for correspondence in large-scale, a variety of self-supervised pretext tasks are proposed to explicitly perform object-level or patch-level similarity learning. Instead of following the previous literature, we propose to learn correspondence using Video Frame-level Similarity (VFS) learning, i.e, simply learning from comparing video frames. Our work is inspired by the recent success in image-level contrastive learning and similarity learning for visual recognition. Our hypothesis is that if the representation is good for recognition, it requires the convolutional features to find correspondence between similar objects or parts. Our experiments show surprising results that VFS surpasses state-of-the-art self-supervised approaches for both OTB visual object tracking and DAVIS video object segmentation. We perform detailed analysis on what matters in VFS and reveals new properties on image and frame level similarity learning. Project page with code is available at https://jerryxu.net/VFS

Jiarui Xu, Xiaolong Wang• 2021

Related benchmarks

TaskDatasetResultRank
Semantic segmentationADE20K (val)
mIoU31.4
2731
Video Object SegmentationDAVIS 2017 (val)
J mean66.5
1130
Semantic segmentationADE20K
mIoU31.4
936
Object DetectionCOCO (val)
mAP41.6
613
Object DetectionLVIS (val)
mAP23.2
141
Visual Object TrackingOTB-100
AUC52.5
136
Object DetectionCOCO
mAP41.6
107
Pose PropagationJHMDB
PCK@0.160.9
20
Video label propagationJHMDB (val)
PCK@0.160.9
17
Human Pose TrackingJHMDB (val)
PCK@.160.5
15
Showing 10 of 23 rows

Other info

Follow for update