Joint Generative and Contrastive Learning for Unsupervised Person Re-identification
About
Recent self-supervised contrastive learning provides an effective approach for unsupervised person re-identification (ReID) by learning invariance from different views (transformed versions) of an input. In this paper, we incorporate a Generative Adversarial Network (GAN) and a contrastive learning module into one joint training framework. While the GAN provides online data augmentation for contrastive learning, the contrastive module learns view-invariant features for generation. In this context, we propose a mesh-based view generator. Specifically, mesh projections serve as references towards generating novel views of a person. In addition, we propose a view-invariant loss to facilitate contrastive learning between original and generated views. Deviating from previous GAN-based unsupervised ReID methods involving domain adaptation, we do not rely on a labeled source dataset, which makes our method more flexible. Extensive experimental results show that our method significantly outperforms state-of-the-art methods under both, fully unsupervised and unsupervised domain adaptive settings on several large scale ReID datsets.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Person Re-Identification | Market1501 (test) | Rank-1 Accuracy87.3 | 1264 | |
| Person Re-Identification | Duke MTMC-reID (test) | Rank-182.9 | 1018 | |
| Person Re-Identification | Market 1501 | mAP66.8 | 999 | |
| Person Re-Identification | MSMT17 (test) | Rank-1 Acc45.7 | 499 | |
| Person Re-Identification | MSMT17 | mAP0.213 | 404 | |
| Person Re-Identification | Market-1501 (test) | Rank-187.3 | 384 | |
| Person Re-Identification | DukeMTMC (test) | mAP62.8 | 83 | |
| Unsupervised Person Re-identification | Market1501 | mAP66.8 | 21 | |
| Unsupervised Person Re-identification | MSMT17 | mAP21.3 | 20 | |
| Person Re-Identification | Market-1501 -> MSMT17 (M→MS) (test) | Rank-151.1 | 7 |