TY - JOUR
T1 - Multi-Projection Fusion and Refinement Network for Salient Object Detection in 360° Omnidirectional Image
AU - CONG, Runmin
AU - HUANG, Ke
AU - LEI, Jianjun
AU - ZHAO, Yao
AU - HUANG, Qingming
AU - KWONG, Sam
N1 - Publisher Copyright:
IEEE
PY - 2023/1/9
Y1 - 2023/1/9
N2 - Salient object detection (SOD) aims to determine the most visually attractive objects in an image. With the development of virtual reality (VR) technology, 360 ∘ omnidirectional image has been widely used, but the SOD task in 360 ∘ omnidirectional image is seldom studied due to its severe distortions and complex scenes. In this article, we propose a multi-projection fusion and refinement network (MPFR-Net) to detect the salient objects in 360 ∘ omnidirectional image. Different from the existing methods, the equirectangular projection (EP) image and four corresponding cube-unfolding (CU) images are embedded into the network simultaneously as inputs, where the CU images not only provide supplementary information for EP image but also ensure the object integrity of cube-map projection. In order to make full use of these two projection modes, a dynamic weighting fusion (DWF) module is designed to adaptively integrate the features of different projections in a complementary and dynamic manner from the perspective of inter and intrafeatures. Furthermore, in order to fully explore the way of interaction between encoder and decoder features, a filtration and refinement (FR) module is designed to suppress the redundant information of the feature itself and between the features. Experimental results on two omnidirectional datasets demonstrate that the proposed approach outperforms the state-of-the-art methods both qualitatively and quantitatively. The code and results can be found from the link of https://rmcong.github.io/proj_MPFRNet.html.
AB - Salient object detection (SOD) aims to determine the most visually attractive objects in an image. With the development of virtual reality (VR) technology, 360 ∘ omnidirectional image has been widely used, but the SOD task in 360 ∘ omnidirectional image is seldom studied due to its severe distortions and complex scenes. In this article, we propose a multi-projection fusion and refinement network (MPFR-Net) to detect the salient objects in 360 ∘ omnidirectional image. Different from the existing methods, the equirectangular projection (EP) image and four corresponding cube-unfolding (CU) images are embedded into the network simultaneously as inputs, where the CU images not only provide supplementary information for EP image but also ensure the object integrity of cube-map projection. In order to make full use of these two projection modes, a dynamic weighting fusion (DWF) module is designed to adaptively integrate the features of different projections in a complementary and dynamic manner from the perspective of inter and intrafeatures. Furthermore, in order to fully explore the way of interaction between encoder and decoder features, a filtration and refinement (FR) module is designed to suppress the redundant information of the feature itself and between the features. Experimental results on two omnidirectional datasets demonstrate that the proposed approach outperforms the state-of-the-art methods both qualitatively and quantitatively. The code and results can be found from the link of https://rmcong.github.io/proj_MPFRNet.html.
KW - 360∘ omnidirectional image
KW - cube-unfolding (CU)
KW - dynamic weighting
KW - filtration and refinement (FR)
KW - salient object detection (SOD)
UR - http://www.scopus.com/inward/record.url?scp=85147315718&partnerID=8YFLogxK
U2 - 10.1109/TNNLS.2022.3233883
DO - 10.1109/TNNLS.2022.3233883
M3 - Journal Article (refereed)
C2 - 37018573
AN - SCOPUS:85147315718
SN - 2162-237X
SP - 1
EP - 13
JO - IEEE Transactions on Neural Networks and Learning Systems
JF - IEEE Transactions on Neural Networks and Learning Systems
ER -