Working hard to know your neighbor's margins: Local descriptor learning loss
About
We introduce a novel loss for learning local feature descriptors which is inspired by the Lowe's matching criterion for SIFT. We show that the proposed loss that maximizes the distance between the closest positive and closest negative patch in the batch is better than complex regularization methods; it works well for both shallow and deep convolution network architectures. Applying the novel loss to the L2Net CNN architecture results in a compact descriptor -- it has the same dimensionality as SIFT (128) that shows state-of-art performance in wide baseline stereo, patch verification and instance retrieval benchmarks. It is fast, computing a descriptor takes about 1 millisecond on a low-end GPU.
Anastasiya Mishchuk, Dmytro Mishkin, Filip Radenovic, Jiri Matas• 2017
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Retrieval | Oxford 5k | mAP88.3 | 100 | |
| Image Retrieval | Oxford5k (test) | mAP85.1 | 97 | |
| Image Retrieval | Paris6k (test) | mAP79.9 | 88 | |
| Image Retrieval | Paris6k | mAP84.9 | 45 | |
| Patch Matching | VIS-NIR (test) | Field Match Rate10.89 | 27 | |
| Patch Matching | UBC Benchmark Liberty, Notre Dame, Yosemite | FPR95 (Train: NOT / Test: LIB)0.53 | 12 | |
| Endoscopic Image Matching | SCARED (test) | Epipolar Error (px)10.45 | 10 | |
| Patch Matching | GAP-VIR (Ground) | FPR@953.94 | 9 | |
| Patch Matching | GAP-VIR Aerial | FPR@952.04 | 9 | |
| Patch Matching | GAP-VIR Combined | FPR953.24 | 3 |
Showing 10 of 10 rows