DAG-Recurrent Neural Networks For Scene Labeling
About
In image labeling, local representations for image units are usually generated from their surrounding image patches, thus long-range contextual information is not effectively encoded. In this paper, we introduce recurrent neural networks (RNNs) to address this issue. Specifically, directed acyclic graph RNNs (DAG-RNNs) are proposed to process DAG-structured images, which enables the network to model long-range semantic dependencies among image units. Our DAG-RNNs are capable of tremendously enhancing the discriminative power of local representations, which significantly benefits the local classification. Meanwhile, we propose a novel class weighting function that attends to rare classes, which phenomenally boosts the recognition accuracy for non-frequent classes. Integrating with convolution and deconvolution layers, our DAG-RNNs achieve new state-of-the-art results on the challenging SiftFlow, CamVid and Barcelona benchmarks.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Scene labeling | CamVid 468/233 (test) | Global Accuracy91.6 | 22 | |
| Semantic segmentation | Barcelona (test) | Global Accuracy74.6 | 18 | |
| Scene labeling | SiftFlow (test) | Global Accuracy85.3 | 15 | |
| Question Answering | bAbI 1.0 (test) | Task 1 Accuracy0.00e+0 | 10 | |
| Video Semantic Segmentation | CamVid (test) | Pixel Accuracy91.6 | 6 |