Person Search via A Mask-Guided Two-Stream CNN Model
About
In this work, we tackle the problem of person search, which is a challenging task consisted of pedestrian detection and person re-identification~(re-ID). Instead of sharing representations in a single joint model, we find that separating detector and re-ID feature extraction yields better performance. In order to extract more representative features for each identity, we segment out the foreground person from the original image patch. We propose a simple yet effective re-ID method, which models foreground person and original image patches individually, and obtains enriched representations from two separate CNN streams. From the experiments on two standard person search benchmarks of CUHK-SYSU and PRW, we achieve mAP of $83.0\%$ and $32.6\%$ respectively, surpassing the state of the art by a large margin (more than 5pp).
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Person Search | CUHK-SYSU (test) | CMC Top-10.837 | 147 | |
| Person Search | PRW (test) | mAP32.6 | 129 | |
| Person Search | CUHK-SYSU v1 (test) | Running Time1.27e+3 | 9 |