Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Bayesian SegNet: Model Uncertainty in Deep Convolutional Encoder-Decoder Architectures for Scene Understanding

About

We present a deep learning framework for probabilistic pixel-wise semantic segmentation, which we term Bayesian SegNet. Semantic segmentation is an important tool for visual scene understanding and a meaningful measure of uncertainty is essential for decision making. Our contribution is a practical system which is able to predict pixel-wise class labels with a measure of model uncertainty. We achieve this by Monte Carlo sampling with dropout at test time to generate a posterior distribution of pixel class labels. In addition, we show that modelling uncertainty improves segmentation performance by 2-3% across a number of state of the art architectures such as SegNet, FCN and Dilation Network, with no additional parametrisation. We also observe a significant improvement in performance for smaller datasets where modelling uncertainty is more effective. We benchmark Bayesian SegNet on the indoor SUN Scene Understanding and outdoor CamVid driving scenes datasets.

Alex Kendall, Vijay Badrinarayanan, Roberto Cipolla• 2015

Related benchmarks

TaskDatasetResultRank
Semantic segmentationCamVid (test)
mIoU63.1
411
Semantic segmentationPascal VOC (test)
mIoU60.5
236
Semantic segmentationNYU Depth V2 (test)
mIoU32.4
172
Semantic segmentationNYUDv2 40-class (test)
mIoU32.4
99
Semantic segmentationNYUD v2
mIoU32.4
96
Semantic segmentationSUN-RGBD (test)
mIoU30.7
77
Semantic segmentationSUN-RGBD 37 classes (test)
mIoU30.7
28
Semantic segmentationNYU Depth V2
Pixel Accuracy68
26
Scene labelingCamVid 468/233 (test)
Global Accuracy86.9
22
Semantic segmentationSUN-RGBD
IoU30.7
19
Showing 10 of 12 rows

Other info

Follow for update