Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Zero-Shot Semantic Segmentation

About

Semantic segmentation models are limited in their ability to scale to large numbers of object classes. In this paper, we introduce the new task of zero-shot semantic segmentation: learning pixel-wise classifiers for never-seen object categories with zero training examples. To this end, we present a novel architecture, ZS3Net, combining a deep visual segmentation model with an approach to generate visual representations from semantic word embeddings. By this way, ZS3Net addresses pixel classification tasks where both seen and unseen categories are faced at test time (so called "generalized" zero-shot classification). Performance is further improved by a self-training step that relies on automatic pseudo-labeling of pixels from unseen classes. On the two standard segmentation datasets, Pascal-VOC and Pascal-Context, we propose zero-shot benchmarks and set competitive baselines. For complex scenes as ones in the Pascal-Context dataset, we extend our approach by using a graph-context encoding to fully leverage spatial context priors coming from class-wise segmentation maps.

Maxime Bucher, Tuan-Hung Vu, Matthieu Cord, Patrick P\'erez• 2019

Related benchmarks

TaskDatasetResultRank
Semantic segmentationADE20K (val)--
2731
Semantic segmentationPASCAL VOC 2012 (val)
Mean IoU61.6
2040
Semantic segmentationPASCAL Context (val)
mIoU26
323
Semantic segmentationPascal VOC (test)
mIoU86.6
236
Semantic segmentationCOCO Stuff
mIoU34.9
195
Semantic segmentationCoco-Stuff (test)
mIoU33.7
184
Few-shot Semantic SegmentationPASCAL-5^i (test)
FB-IoU57.7
177
Semantic segmentationPascal Context (test)--
176
Semantic segmentationPascal VOC
mIoU0.78
172
Semantic segmentationPascal Context 59
mIoU19.4
164
Showing 10 of 52 rows

Other info

Code

Follow for update