Boundary-aware Self-supervised Learning for Video Scene Segmentation
About
Self-supervised learning has drawn attention through its effectiveness in learning in-domain representations with no ground-truth annotations; in particular, it is shown that properly designed pretext tasks (e.g., contrastive prediction task) bring significant performance gains for downstream tasks (e.g., classification task). Inspired from this, we tackle video scene segmentation, which is a task of temporally localizing scene boundaries in a video, with a self-supervised learning framework where we mainly focus on designing effective pretext tasks. In our framework, we discover a pseudo-boundary from a sequence of shots by splitting it into two continuous, non-overlapping sub-sequences and leverage the pseudo-boundary to facilitate the pre-training. Based on this, we introduce three novel boundary-aware pretext tasks: 1) Shot-Scene Matching (SSM), 2) Contextual Group Matching (CGM) and 3) Pseudo-boundary Prediction (PP); SSM and CGM guide the model to maximize intra-scene similarity and inter-scene discrimination while PP encourages the model to identify transitional moments. Through comprehensive analysis, we empirically show that pre-training and transferring contextual representation are both critical to improving the video scene segmentation performance. Lastly, we achieve the new state-of-the-art on the MovieNet-SSeg benchmark. The code is available at https://github.com/kakaobrain/bassl.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Movie Scene Segmentation | MovieNet 1.0 (test) | AP57.4 | 20 | |
| Scene Detection | OVSD | AP28.68 | 8 | |
| Dialogue Scene Segmentation | VSTAR (test) | mIoU46.6 | 7 | |
| Scene Segmentation | MovieNet-318 (test) | AP57.4 | 6 | |
| Scene Detection | BBC | AP0.3998 | 5 | |
| Scene Segmentation | BBC Planet Earth (test) | AP0.4 | 4 |