A Deep Multi-Level Network for Saliency Prediction
About
This paper presents a novel deep architecture for saliency prediction. Current state of the art models for saliency prediction employ Fully Convolutional networks that perform a non-linear combination of features extracted from the last convolutional layer to predict saliency maps. We propose an architecture which, instead, combines features extracted at different levels of a Convolutional Neural Network (CNN). Our model is composed of three main blocks: a feature extraction CNN, a feature encoding network, that weights low and high level feature maps, and a prior learning network. We compare our solution with state of the art saliency models on two public benchmarks datasets. Results show that our model outperforms under all evaluation metrics on the SALICON dataset, which is currently the largest public dataset for saliency prediction, and achieves competitive results on the MIT300 benchmark.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Saliency Prediction | MIT300 (test) | CC0.67 | 56 | |
| Visual Saliency Prediction | CAT2000 (test) | Correlation Coefficient (CC)0.5221 | 19 | |
| Saliency Prediction | MIT1003 (test) | NSS2.3329 | 18 | |
| Distortion-aware saliency prediction | GenBlemish-27K | AUC-Judd0.8539 | 17 | |
| Visual Saliency Prediction | SALICON (test) | CC0.743 | 12 | |
| Affordance Grounding | OPRA 28 x 28 (test) | KLD4.02 | 11 | |
| Affordance Grounding | EPIC-Hotspots 28 x 28 (test) | KLD6.12 | 10 | |
| Grounded affordance prediction | OPRA (seen classes) | KLD4.022 | 9 | |
| Affordance Grounding | OPRA (test) | KLD4.022 | 9 | |
| Generalization to novel objects | OPRA novel objects | KLD2.458 | 8 |