A weakly supervised adaptive triplet loss for deep metric learning
About
We address the problem of distance metric learning in visual similarity search, defined as learning an image embedding model which projects images into Euclidean space where semantically and visually similar images are closer and dissimilar images are further from one another. We present a weakly supervised adaptive triplet loss (ATL) capable of capturing fine-grained semantic similarity that encourages the learned image embedding models to generalize well on cross-domain data. The method uses weakly labeled product description data to implicitly determine fine grained semantic classes, avoiding the need to annotate large amounts of training data. We evaluate on the Amazon fashion retrieval benchmark and DeepFashion in-shop retrieval data. The method boosts the performance of triplet loss baseline by 10.6% on cross-domain data and out-performs the state-of-art model on all evaluation metrics.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Medical Image Retrieval | CXR 1 year (test) | mAP81.5 | 7 | |
| Medical Image Retrieval | CXR 2 years (test) | mAP72.7 | 7 | |
| Medical Image Retrieval | CXR 3 years (test) | mAP66.1 | 7 | |
| Medical Image Retrieval | CXR 4 years (test) | mAP67.1 | 7 | |
| Medical Image Retrieval | CXR 10 years (test) | mAP69.9 | 7 | |
| Medical Image Retrieval | CXR 11 years (test) | mAP0.747 | 7 | |
| Medical Image Retrieval | CXR 12 years (test) | mAP0.579 | 7 | |
| Medical Image Retrieval | CXR All (test) | mAP73.2 | 7 | |
| Medical Image Retrieval | CXR 6 years (test) | mAP67.1 | 7 | |
| Medical Image Retrieval | CXR 7 years (test) | mAP43.1 | 7 |