Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

STAViS: Spatio-Temporal AudioVisual Saliency Network

About

We introduce STAViS, a spatio-temporal audiovisual saliency network that combines spatio-temporal visual and auditory information in order to efficiently address the problem of saliency estimation in videos. Our approach employs a single network that combines visual saliency and auditory features and learns to appropriately localize sound sources and to fuse the two saliencies in order to obtain a final saliency map. The network has been designed, trained end-to-end, and evaluated on six different databases that contain audiovisual eye-tracking data of a large variety of videos. We compare our method against 8 different state-of-the-art visual saliency models. Evaluation results across databases indicate that our STAViS model outperforms our visual only variant as well as the other state-of-the-art models in the majority of cases. Also, the consistently good performance it achieves for all databases indicates that it is appropriate for estimating saliency "in-the-wild". The code is available at https://github.com/atsiami/STAViS.

Antigoni Tsiami, Petros Koutras, Petros Maragos• 2020

Related benchmarks

TaskDatasetResultRank
Saliency PredictionDIEM (test)
SIM0.482
28
Video saliency predictionETMD (test)
CC0.569
7
Video saliency predictionCoutrot2 (test)
CC0.734
7
Video saliency predictionAVAD (test)
CC0.608
7
Video saliency predictionCoutrot1 (test)
CC0.472
7
Video saliency predictionSumMe (test)
CC0.422
7
Video saliency predictionSTA (test)
AUC-J0.8752
3
Showing 7 of 7 rows

Other info

Follow for update