Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

TransforMatcher: Match-to-Match Attention for Semantic Correspondence

About

Establishing correspondences between images remains a challenging task, especially under large appearance changes due to different viewpoints or intra-class variations. In this work, we introduce a strong semantic image matching learner, dubbed TransforMatcher, which builds on the success of transformer networks in vision domains. Unlike existing convolution- or attention-based schemes for correspondence, TransforMatcher performs global match-to-match attention for precise match localization and dynamic refinement. To handle a large number of matches in a dense correlation map, we develop a light-weight attention architecture to consider the global match-to-match interactions. We also propose to utilize a multi-channel correlation map for refinement, treating the multi-level scores as features instead of a single score to fully exploit the richer layer-wise semantics. In experiments, TransforMatcher sets a new state of the art on SPair-71k while performing on par with existing SOTA methods on the PF-PASCAL dataset.

Seungwook Kim, Juhong Min, Minsu Cho• 2022

Related benchmarks

TaskDatasetResultRank
Semantic CorrespondenceSPair-71k (test)
PCK@0.157.9
122
Semantic CorrespondencePF-WILLOW
PCK@0.1 (bbox)76
109
Semantic CorrespondencePF-Pascal (test)
PCK@0.193.3
106
Semantic CorrespondencePF-PASCAL
PCK @ alpha=0.191.8
98
Semantic CorrespondencePF-WILLOW (test)--
37
Semantic CorrespondenceSPair-71k
Φ_bbox @ α=0.130.1
29
Semantic MatchingSPair-71k 1.0 (test)
PCK@0.1 (Aero)59.2
16
Semantic MatchingSPair-71k (val)
PCK (@alpha_bbox=0.1)53.7
15
Showing 8 of 8 rows

Other info

Code

Follow for update