Use All The Labels: A Hierarchical Multi-Label Contrastive Learning Framework
About
Current contrastive learning frameworks focus on leveraging a single supervisory signal to learn representations, which limits the efficacy on unseen data and downstream tasks. In this paper, we present a hierarchical multi-label representation learning framework that can leverage all available labels and preserve the hierarchical relationship between classes. We introduce novel hierarchy preserving losses, which jointly apply a hierarchical penalty to the contrastive loss, and enforce the hierarchy constraint. The loss function is data driven and automatically adapts to arbitrary multi-label structures. Experiments on several datasets show that our relationship-preserving embedding performs well on a variety of tasks and outperform the baseline supervised and self-supervised approaches. Code is available at https://github.com/salesforce/hierarchicalContrastiveLearning.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| In-shop clothing retrieval | DeepFashion in-shop | -- | 26 | |
| Image Classification | DeepFashion In-Shop (Unseen) | Top-1 Acc75.29 | 7 | |
| Image Classification | ModelNet40 (Seen) | Top-1 Accuracy89.45 | 7 | |
| Image Classification | ModelNet40 Unseen | Top-1 Accuracy85.37 | 7 | |
| Image Classification | DeepFashion In-Shop (Seen) | Top-1 Accuracy80.67 | 7 | |
| Image Retrieval | ModelNet40 (test) | mAP26 | 7 | |
| Clustering | DeepFashion in-shop | Category NMI59 | 6 | |
| Clustering | ModelNet40 | NMI (Category)0.62 | 6 | |
| Image Classification | ImageNet finest sub-category level (full) | Top-1 Acc79.14 | 5 | |
| Image Classification | DeepFashion super-category level (full) | Top-1 Acc73.21 | 5 |