Transferring a Semantic Representation for Person Re-Identification and Search
About
Learning semantic attributes for person re-identification and description-based person search has gained increasing interest due to attributes' great potential as a pose and view-invariant representation. However, existing attribute-centric approaches have thus far underperformed state-of-the-art conventional approaches. This is due to their non-scalable need for extensive domain (camera) specific annotation. In this paper we present a new semantic attribute learning approach for person re-identification and search. Our model is trained on existing fashion photography datasets -- either weakly or strongly labelled. It can then be transferred and adapted to provide a powerful semantic description of surveillance person detections, without requiring any surveillance domain supervision. The resulting representation is useful for both unsupervised and supervised person re-identification, achieving state-of-the-art and near state-of-the-art performance respectively. Furthermore, as a semantic representation it allows description-based person search to be integrated within the same framework.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Person Re-Identification | VIPeR | Rank-141.6 | 182 | |
| Person Re-Identification | VIPeR (test) | Top-1 Accuracy41.6 | 113 | |
| Person Re-Identification | CUHK01 (486/485 split) | R131.5 | 30 | |
| Person Re-Identification | PRID450S | Rank-1 Acc44.9 | 8 |