Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

COTR: Correspondence Transformer for Matching Across Images

About

We propose a novel framework for finding correspondences in images based on a deep neural network that, given two images and a query point in one of them, finds its correspondence in the other. By doing so, one has the option to query only the points of interest and retrieve sparse correspondences, or to query all points in an image and obtain dense mappings. Importantly, in order to capture both local and global priors, and to let our model relate between image regions using the most relevant among said priors, we realize our network using a transformer. At inference time, we apply our correspondence network by recursively zooming in around the estimates, yielding a multiscale pipeline able to provide highly-accurate correspondences. Our method significantly outperforms the state of the art on both sparse and dense correspondence problems on multiple datasets and tasks, ranging from wide-baseline stereo to optical flow, without any retraining for a specific dataset. We commit to releasing data, code, and all the tools necessary to train from scratch and ensure reproducibility.

Wei Jiang, Eduard Trulls, Jan Hosang, Andrea Tagliasacchi, Kwang Moo Yi• 2021

Related benchmarks

TaskDatasetResultRank
Optical Flow EstimationKITTI 2015 (test)
Fl-all13.65
91
Point TrackingDAVIS TAP-Vid
Average Jaccard (AJ)35.4
41
Point TrackingDAVIS
AJ35.4
38
Point TrackingTAP-Vid Kinetics
Overall Accuracy57.4
37
Homography EstimationHPatches
AUC @3px41.9
35
Point TrackingTAP-Vid RGB-Stacking (test)
AJ6.8
32
Point TrackingTAP-Vid DAVIS (test)
AJ35.4
31
Point TrackingTAP-Vid Kinetics (test)
Average Jitter (AJ)19
30
Point TrackingKinetics
delta_avg38.8
24
Point TrackingDAVIS TAP-Vid (val)
AJ35.4
19
Showing 10 of 23 rows

Other info

Follow for update