SalBiNet360: Saliency Prediction on 360° Images with Local-Global Bifurcated Deep Network

Note: We don't have the ability to review paper

PubDate: May 2020

Teams: South China University of Technology

Writers: Dongwen Chen; Chunmei Qing; Xiangmin Xu; Huansheng Zhu

PDF: SalBiNet360: Saliency Prediction on 360° Images with Local-Global Bifurcated Deep Network

Abstract

With the development of the virtual reality applications, predicting human visual attention on 360° images is valuable to content creators and encoding algorithms, and becomes essential to understand user behaviour. In this paper, we propose a local-global bifurcated deep network for saliency prediction on 360° images, which is named as SalBiNet360. In the global deep sub-network, multiple multi-scale contextual modules and a multilevel decoder are utilized to integrate the features from the middle and deep layers of the network. In the local deep sub-network, only one multi-scale contextual module and a single-level decoder are utilized to reduce the redundancy of local saliency maps. Finally, fused saliency maps are generated by linear combination of the global and local saliency maps. Experiments on two publicly available datasets illustrate that the proposed SalBiNet360 outperforms the tested state-of-the-art methods.

更多阅读推荐......

Paper