Zhang Miao, Xu Shuang, Piao Yongri, Lu Huchuan
IEEE Trans Image Process. 2022;31:6152-6163. doi: 10.1109/TIP.2022.3205749. Epub 2022 Sep 22.
Previous 2D saliency detection methods extract salient cues from a single view and directly predict the expected results. Both traditional and deep-learning-based 2D methods do not consider geometric information of 3D scenes. Therefore the relationship between scene understanding and salient objects cannot be effectively established. This limits the performance of 2D saliency detection in challenging scenes. In this paper, we show for the first time that saliency detection problem can be reformulated as two sub-problems: light field synthesis from a single view and light-field-driven saliency detection. This paper first introduces a high-quality light field synthesis network to produce reliable 4D light field information. Then a novel light-field-driven saliency detection network is proposed, in which a Direction-specific Screening Unit (DSU) is tailored to exploit the spatial correlation among multiple viewpoints. The whole pipeline can be trained in an end-to-end fashion. Experimental results demonstrate that the proposed method outperforms the state-of-the-art 2D, 3D and 4D saliency detection methods. Our code is publicly available at https://github.com/OIPLab-DUT/ESCNet.
以往的二维显著性检测方法从单一视角提取显著线索并直接预测预期结果。基于传统方法和深度学习的二维方法都没有考虑三维场景的几何信息。因此,无法有效建立场景理解与显著物体之间的关系。这限制了二维显著性检测在具有挑战性场景中的性能。在本文中,我们首次表明显著性检测问题可以重新表述为两个子问题:从单视角进行光场合成和光场驱动的显著性检测。本文首先引入一个高质量的光场合成网络来生成可靠的四维光场信息。然后提出了一种新颖的光场驱动的显著性检测网络,其中定制了一个方向特定筛选单元(DSU)来利用多个视角之间的空间相关性。整个流程可以以端到端的方式进行训练。实验结果表明,所提出的方法优于当前最先进的二维、三维和四维显著性检测方法。我们的代码可在https://github.com/OIPLab-DUT/ESCNet上公开获取。