Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Three Ways to Improve Semantic Segmentation with Self-Supervised Depth Estimation

About

Training deep networks for semantic segmentation requires large amounts of labeled training data, which presents a major challenge in practice, as labeling segmentation masks is a highly labor-intensive process. To address this issue, we present a framework for semi-supervised semantic segmentation, which is enhanced by self-supervised monocular depth estimation from unlabeled image sequences. In particular, we propose three key contributions: (1) We transfer knowledge from features learned during self-supervised depth estimation to semantic segmentation, (2) we implement a strong data augmentation by blending images and labels using the geometry of the scene, and (3) we utilize the depth feature diversity as well as the level of difficulty of learning depth in a student-teacher framework to select the most useful samples to be annotated for semantic segmentation. We validate the proposed model on the Cityscapes dataset, where all three modules demonstrate significant performance gains, and we achieve state-of-the-art results for semi-supervised semantic segmentation. The implementation is available at https://github.com/lhoyer/improving_segmentation_with_selfsupervised_depth.

Lukas Hoyer, Dengxin Dai, Yuhua Chen, Adrian K\"oring, Suman Saha, Luc Van Gool• 2020

Related benchmarks

TaskDatasetResultRank
Semantic segmentationCityscapes (test)
mIoU75
1145
Depth EstimationNYU v2 (test)--
423
Surface Normal EstimationNYU v2 (test)--
206
Semantic segmentationNYUD v2 (test)
mIoU40.28
187
Semantic segmentationNYU Depth V2 (test)
mIoU39.47
172
Depth PredictionCityscapes (test)
RMSE6.528
52
Multi-task LearningCityscapes (test)
MR41.84
43
Edge DetectionNYUD v2 (test)--
16
Multi-task LearningSynthia (test)
mIoU79.93
10
Multi-task LearningvKITTI 2 (test)
mIoU96.87
10
Showing 10 of 11 rows

Other info

Code

Follow for update