Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

SCAN: Self-and-Collaborative Attention Network for Video Person Re-identification

About

Video person re-identification attracts much attention in recent years. It aims to match image sequences of pedestrians from different camera views. Previous approaches usually improve this task from three aspects, including a) selecting more discriminative frames, b) generating more informative temporal representations, and c) developing more effective distance metrics. To address the above issues, we present a novel and practical deep architecture for video person re-identification termed Self-and-Collaborative Attention Network (SCAN). It has several appealing properties. First, SCAN adopts non-parametric attention mechanism to refine the intra-sequence and inter-sequence feature representation of videos, and outputs self-and-collaborative feature representation for each video, making the discriminative frames aligned between the probe and gallery sequences.Second, beyond existing models, a generalized pairwise similarity measurement is proposed to calculate the similarity feature representations of video pairs, enabling computing the matching scores by the binary classifier. Third, a dense clip segmentation strategy is also introduced to generate rich probe-gallery pairs to optimize the model. Extensive experiments demonstrate the effectiveness of SCAN, which outperforms the best-performing baselines on iLIDS-VID, PRID2011 and MARS dataset, respectively.

Ruimao Zhang, Hongbin Sun, Jingyu Li, Yuying Ge, Liang Lin, Ping Luo, Xiaogang Wang• 2018

Related benchmarks

TaskDatasetResultRank
Person Re-IdentificationiLIDS-VID
CMC-188
80
Person Re-IdentificationMARS (test)
Rank-187.2
72
Person Re-IdentificationPRID2011
Rank-195.3
66
Showing 3 of 3 rows

Other info

Follow for update