A Strong Baseline for Fashion Retrieval with Person Re-Identification Models
About
Fashion retrieval is the challenging task of finding an exact match for fashion items contained within an image. Difficulties arise from the fine-grained nature of clothing items, very large intra-class and inter-class variance. Additionally, query and source images for the task usually come from different domains - street photos and catalogue photos respectively. Due to these differences, a significant gap in quality, lighting, contrast, background clutter and item presentation exists between domains. As a result, fashion retrieval is an active field of research both in academia and the industry. Inspired by recent advancements in Person Re-Identification research, we adapt leading ReID models to be used in fashion retrieval tasks. We introduce a simple baseline model for fashion retrieval, significantly outperforming previous state-of-the-art results despite a much simpler architecture. We conduct in-depth experiments on Street2Shop and DeepFashion datasets and validate our results. Finally, we propose a cross-domain (cross-dataset) evaluation method to test the robustness of fashion retrieval models.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Person Re-Identification | Market1501 (test) | Rank-1 Accuracy98 | 1264 | |
| Person Re-Identification | Duke MTMC-reID (test) | Rank-194.5 | 1018 | |
| In-shop clothing retrieval | DeepFashion in-shop | Top-1 Accuracy37.8 | 26 | |
| Fashion Retrieval | Street2Shop | mAP46.8 | 6 | |
| Fashion Retrieval | DeepFashion | Acc@140 | 3 | |
| Fashion Retrieval | Street2Shop (test) | mAP37.2 | 3 | |
| Fashion Retrieval | DeepFashion (test) | Acc@130.8 | 3 |