Learning visual groups from co-occurrences in space and time
About
We propose a self-supervised framework that learns to group visual entities based on their rate of co-occurrence in space and time. To model statistical dependencies between the entities, we set up a simple binary classification problem in which the goal is to predict if two visual primitives occur in the same spatial or temporal context. We apply this framework to three domains: learning patch affinities from spatial adjacency in images, learning frame affinities from temporal adjacency in videos, and learning photo affinities from geospatial proximity in image collections. We demonstrate that in each case the learned affinities uncover meaningful semantic groupings. From patch affinities we generate object proposals that are competitive with state-of-the-art supervised methods. From frame affinities we generate movie scene segmentations that correlate well with DVD chapter structure. Finally, from geospatial affinities we learn groups that relate well to semantic place categories.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Semantic segmentation | Cityscapes | -- | 578 | |
| Semantic segmentation | PASCAL VOC (val) | mIoU400 | 338 | |
| Semantic segmentation | COCO Stuff | -- | 195 | |
| Semantic segmentation | Coco-Stuff (test) | -- | 184 | |
| Semantic segmentation | Potsdam (test) | -- | 104 | |
| Semantic segmentation | COCO Stuff-27 (val) | -- | 75 | |
| Semantic segmentation | COCO-Stuff 27 | -- | 40 | |
| Unsupervised image segmentation | Coco-Stuff (test) | Accuracy24.3 | 26 | |
| Semantic segmentation | Potsdam-3 | Pixel Accuracy63.9 | 25 | |
| Semantic segmentation | PASCAL (val) | mIoU13.5 | 25 |