Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

LinkNet: Exploiting Encoder Representations for Efficient Semantic Segmentation

About

Pixel-wise semantic segmentation for visual scene understanding not only needs to be accurate, but also efficient in order to find any use in real-time application. Existing algorithms even though are accurate but they do not focus on utilizing the parameters of neural network efficiently. As a result they are huge in terms of parameters and number of operations; hence slow too. In this paper, we propose a novel deep neural network architecture which allows it to learn without any significant increase in number of parameters. Our network uses only 11.5 million parameters and 21.2 GFLOPs for processing an image of resolution 3x640x360. It gives state-of-the-art performance on CamVid and comparable results on Cityscapes dataset. We also compare our networks processing time on NVIDIA GPU and embedded system device with existing state-of-the-art architectures for different image resolutions.

Abhishek Chaurasia, Eugenio Culurciello• 2017

Related benchmarks

TaskDatasetResultRank
Semantic segmentationCityscapes (test)--
1145
Semantic segmentationCityscapes (val)--
572
Semantic segmentationLoveDA (test)
mIoU48.5
81
Semantic segmentationMapillary Vistas (val)
mIoU47.7
72
Semantic segmentationLoveDA
IoU (Background)43.61
60
Semantic segmentationCityscapes
Throughput (FPS)65.8
42
Road SegmentationMassachusetts Road Dataset
IoU (Average)0.6312
35
Semantic segmentationPASS (test)
mIoU30.5
28
Semantic segmentationEarthVLSet (test)
mIoU51.02
21
Road SegmentationDeepGlobe
Precision78.34
18
Showing 10 of 16 rows

Other info

Follow for update