SOLAR: Second-Order Loss and Attention for Image Retrieval
About
Recent works in deep-learning have shown that second-order information is beneficial in many computer-vision tasks. Second-order information can be enforced both in the spatial context and the abstract feature dimensions. In this work, we explore two second-order components. One is focused on second-order spatial information to increase the performance of image descriptors, both local and global. It is used to re-weight feature maps, and thus emphasise salient image locations that are subsequently used for description. The second component is concerned with a second-order similarity (SOS) loss, that we extend to global descriptors for image retrieval, and is used to enhance the triplet loss with hard-negative mining. We validate our approach on two different tasks and datasets for image retrieval and image matching. The results show that our two second-order components complement each other, bringing significant performance improvements in both tasks and lead to state-of-the-art results across the public benchmarks. Code available at: http://github.com/tonyngjichun/SOLAR
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Retrieval | RParis Revisited (medium, hard) | mAP (medium)81.6 | 9 | |
| Image Retrieval | RParis + R1M medium hard | mAP (medium)59.2 | 9 | |
| Image Retrieval | ROxford Revisited (medium, hard) | mAP (medium)69.9 | 9 | |
| Image Retrieval | ROxford + R1M (medium, hard) | mAP (medium)53.5 | 9 |