DeepCut: Object Segmentation from Bounding Box Annotations using Convolutional Neural Networks
About
In this paper, we propose DeepCut, a method to obtain pixelwise object segmentations given an image dataset labelled with bounding box annotations. It extends the approach of the well-known GrabCut method to include machine learning by training a neural network classifier from bounding box annotations. We formulate the problem as an energy minimisation problem over a densely-connected conditional random field and iteratively update the training targets to obtain pixelwise object segmentations. Additionally, we propose variants of the DeepCut method and compare those to a naive approach to CNN training under weak supervision. We test its applicability to solve brain and lung segmentation problems on a challenging fetal magnetic resonance dataset and obtain encouraging results in terms of accuracy.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Vestibular Schwannoma Segmentation | Vestibular Schwannoma (VS) dataset (test) | Dice52.4 | 9 | |
| Prostate Segmentation | PROMISE 2012 (test) | -- | 6 | |
| Image Segmentation | ATLAS (test) | Dice Coefficient37.5 | 5 |