Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Recurrent Transformer Networks for Semantic Correspondence

About

We present recurrent transformer networks (RTNs) for obtaining dense correspondences between semantically similar images. Our networks accomplish this through an iterative process of estimating spatial transformations between the input images and using these transformations to generate aligned convolutional activations. By directly estimating the transformations between an image pair, rather than employing spatial transformer networks to independently normalize each individual image, we show that greater accuracy can be achieved. This process is conducted in a recursive manner to refine both the transformation estimates and the feature representations. In addition, a technique is presented for weakly-supervised training of RTNs that is based on a proposed classification loss. With RTNs, state-of-the-art performance is attained on several benchmarks for semantic correspondence.

Seungryong Kim, Stephen Lin, Sangryul Jeon, Dongbo Min, Kwanghoon Sohn• 2018

Related benchmarks

TaskDatasetResultRank
Semantic CorrespondenceSPair-71k (test)
PCK@0.125.7
122
Semantic CorrespondencePF-WILLOW
PCK@0.1 (bbox)71.9
109
Semantic CorrespondencePF-Pascal (test)
PCK@0.175.9
106
Semantic CorrespondencePF-PASCAL
PCK @ alpha=0.175.9
98
Semantic CorrespondencePF-WILLOW (test)--
37
Semantic keypoint transferPF-Pascal (test)
PCK @ 0.0555.2
35
Semantic CorrespondencePF-PASCAL
PCK@0.175.9
29
Semantic CorrespondenceSPair-71k
PCK@0.125.7
24
Semantic CorrespondencePF-PASCAL (val)
PCK @ 0.0555.2
8
Showing 9 of 9 rows

Other info

Follow for update