Depth-aware CNN for RGB-D Segmentation
About
Convolutional neural networks (CNN) are limited by the lack of capability to handle geometric information due to the fixed grid kernel structure. The availability of depth data enables progress in RGB-D semantic segmentation with CNNs. State-of-the-art methods either use depth as additional images or process spatial information in 3D volumes or point clouds. These methods suffer from high computation and memory cost. To address these issues, we present Depth-aware CNN by introducing two intuitive, flexible and effective operations: depth-aware convolution and depth-aware average pooling. By leveraging depth similarity between pixels in the process of information propagation, geometry is seamlessly incorporated into CNN. Without introducing any additional parameters, both operators can be easily integrated into existing CNNs. Extensive experiments and ablation studies on challenging RGB-D semantic segmentation benchmarks validate the effectiveness and flexibility of our approach.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Semantic segmentation | NYU v2 (test) | mIoU43.9 | 248 | |
| Semantic segmentation | SUN RGB-D (test) | mIoU42 | 191 | |
| Semantic segmentation | NYUD v2 (test) | mIoU48.4 | 187 | |
| Semantic segmentation | NYU Depth V2 (test) | mIoU43.9 | 172 | |
| Semantic segmentation | NYUDv2 40-class (test) | mIoU48.4 | 99 | |
| Semantic segmentation | SUN-RGBD (test) | mIoU42 | 77 | |
| Semantic segmentation | MFNet nighttime (test) | mIoU43.2 | 42 | |
| Scene Parsing | NYUDv2 (test) | mIoU43.9 | 35 | |
| Semantic segmentation | MFNet daytime (test) | mIoU42.4 | 30 | |
| Semantic segmentation | SUN-RGBD 37 classes (test) | mIoU42 | 28 |