Fully Convolutional Networks for Panoptic Segmentation
About
In this paper, we present a conceptually simple, strong, and efficient framework for panoptic segmentation, called Panoptic FCN. Our approach aims to represent and predict foreground things and background stuff in a unified fully convolutional pipeline. In particular, Panoptic FCN encodes each object instance or stuff category into a specific kernel weight with the proposed kernel generator and produces the prediction by convolving the high-resolution feature directly. With this approach, instance-aware and semantically consistent properties for things and stuff can be respectively satisfied in a simple generate-kernel-then-segment workflow. Without extra boxes for localization or instance separation, the proposed approach outperforms previous box-based and -free models with high efficiency on COCO, Cityscapes, and Mapillary Vistas datasets with single scale input. Our code is made publicly available at https://github.com/Jia-Research-Lab/PanopticFCN.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Panoptic Segmentation | Cityscapes (val) | PQ61.4 | 276 | |
| Panoptic Segmentation | COCO (val) | PQ51.8 | 219 | |
| Panoptic Segmentation | COCO 2017 (val) | PQ43.6 | 172 | |
| Panoptic Segmentation | COCO (test-dev) | PQ47.5 | 162 | |
| Panoptic Segmentation | Mapillary Vistas (val) | PQ36.9 | 82 | |
| Panoptic Segmentation | COCO 2017 (test-dev) | PQ45.5 | 41 | |
| Panoptic Segmentation | COCO (test) | PQ41 | 23 | |
| Panoptic Segmentation | COCO 2000 image subset 2017 (val) | PQ43.8 | 7 | |
| Panoptic Segmentation | COCO-C 2000 image subset | Mean PQ26.8 | 7 |