Deep Shape Matching
About
We cast shape matching as metric learning with convolutional networks. We break the end-to-end process of image representation into two parts. Firstly, well established efficient methods are chosen to turn the images into edge maps. Secondly, the network is trained with edge maps of landmark images, which are automatically obtained by a structure-from-motion pipeline. The learned representation is evaluated on a range of different tasks, providing improvements on challenging cases of domain generalization, generic sketch-based image retrieval or its fine-grained counterpart. In contrast to other methods that learn a different model per task, object category, or domain, we use the same network throughout all our experiments, achieving state-of-the-art results in multiple benchmarks.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Multi-class classification | PACS (test) | Accuracy (Art Painting)80 | 76 | |
| Image Retrieval | Paris Revisited (Medium) | mAP43.5 | 63 | |
| Image Retrieval | R-Oxford Medium | mAP17.3 | 35 | |
| Image Retrieval | Tokyo 24/7 (test) | mAP75.9 | 34 | |
| Sketch-based image retrieval | Flickr15k (test) | mAP68.9 | 17 | |
| Visual Place Recognition | Tokyo 24/7 (test) | mAP75.9 | 13 | |
| Sketch-based image retrieval | Chairs (test) | Top-1 Acc85.6 | 9 | |
| Sketch-based image retrieval | Handbags (test) | Acc@151.2 | 9 | |
| Sketch-based image retrieval | Shoes (test) | Acc@10.548 | 9 |