Chen Chenglizhao, Wei Jipeng, Peng Chong, Zhang Weizhong, Qin Hong, University Qingdao, University Stony Brook
IEEE Trans Image Process. 2020 Jan 30. doi: 10.1109/TIP.2020.2968250.
To solve the saliency detection problem in RGB-D images, the depth information plays a critical role in distinguishing salient objects or foregrounds from cluttered backgrounds. As the complementary component to color information, the depth quality directly dictates the subsequent saliency detection performance. However, due to artifacts and the limitation of depth acquisition devices, the quality of the obtained depth varies tremendously across different scenarios. Consequently, conventional selective fusion-based RGB-D saliency detection methods may result in a degraded detection performance in cases containing salient objects with low color contrast coupled with a low depth quality. To solve this problem, we make our initial attempt to estimate additional high-quality depth information, which is denoted by Depth+. Serving as a complement to the original depth, Depth+ will be fed into our newly designed selective fusion network to boost the detection performance. To achieve this aim, we first retrieve a small group of images that are similar to the given input, and then the inter-image, nonlocal correspondences are built accordingly. Thus, by using these inter-image correspondences, the overall depth can be coarsely estimated by utilizing our newly designed depth-transferring strategy. Next, we build fine-grained, object-level correspondences coupled with a saliency prior to further improve the depth quality of the previous estimation. Compared to the original depth, our newly estimated Depth+ is potentially more informative for detection improvement. Finally, we feed both the original depth and the newly estimated Depth+ into our selective deep fusion network, whose key novelty is to achieve an optimal complementary balance to make better decisions toward improving saliency boundaries.
为了解决RGB-D图像中的显著性检测问题,深度信息在从杂乱背景中区分显著物体或前景方面起着关键作用。作为颜色信息的补充成分,深度质量直接决定了后续的显著性检测性能。然而,由于伪影和深度采集设备的限制,在不同场景下获取的深度质量差异极大。因此,传统的基于选择性融合的RGB-D显著性检测方法在包含颜色对比度低且深度质量差的显著物体的情况下,可能会导致检测性能下降。为了解决这个问题,我们首次尝试估计额外的高质量深度信息,记为Depth+。作为原始深度的补充,Depth+将被输入到我们新设计的选择性融合网络中以提高检测性能。为了实现这一目标,我们首先检索一小部分与给定输入相似的图像,然后相应地建立图像间的非局部对应关系。因此,通过使用这些图像间对应关系,可以利用我们新设计的深度传递策略粗略估计整体深度。接下来,我们结合显著性先验建立细粒度的物体级对应关系,以进一步提高先前估计的深度质量。与原始深度相比,我们新估计的Depth+在改进检测方面可能具有更多信息。最后,我们将原始深度和新估计的Depth+都输入到我们的选择性深度融合网络中,其关键新颖之处在于实现最佳的互补平衡,以便在改进显著性边界方面做出更好的决策。