Suppr超能文献

用于 RGB-D 显著目标检测的判别式跨模态迁移学习和密集跨层反馈融合。

Discriminative Cross-Modal Transfer Learning and Densely Cross-Level Feedback Fusion for RGB-D Salient Object Detection.

出版信息

IEEE Trans Cybern. 2020 Nov;50(11):4808-4820. doi: 10.1109/TCYB.2019.2934986. Epub 2019 Aug 30.

Abstract

This article addresses two key issues in RGB-D salient object detection based on the convolutional neural network (CNN). 1) How to bridge the gap between the "data-hungry" nature of CNNs and the insufficient labeled training data in the depth modality? 2) How to take full advantages of the complementary information among two modalities. To solve the first problem, we model the depth-induced saliency detection as a CNN-based cross-modal transfer learning problem. Instead of directly adopting the RGB CNN as initialization, we additionally train a modality classification network (MCNet) to encourage discriminative modality-specific representations in minimizing the modality classification loss. To solve the second problem, we propose a densely cross-level feedback topology, in which the cross-modal complements are combined in each level and then densely fed back to all shallower layers for sufficient cross-level interactions. Compared to traditional two-stream frameworks, the proposed one can better explore, select, and fuse cross-modal cross-level complements. Experiments show the significant and consistent improvements of the proposed CNN framework over other state-of-the-art methods.

摘要

本文针对基于卷积神经网络(CNN)的 RGB-D 显著目标检测中的两个关键问题。1)如何弥合 CNN 的“数据饥渴”性质与深度模态中不足的标记训练数据之间的差距?2)如何充分利用两种模态之间的互补信息。为了解决第一个问题,我们将深度诱导的显著检测建模为基于 CNN 的跨模态迁移学习问题。我们不是直接采用 RGB CNN 作为初始化,而是额外训练一个模态分类网络(MCNet),通过最小化模态分类损失来鼓励具有判别力的模态特定表示。为了解决第二个问题,我们提出了一种密集的跨层反馈拓扑结构,其中在每个级别中组合跨模态互补,并将其密集地反馈到所有较浅层,以实现充分的跨层交互。与传统的双流框架相比,所提出的方法可以更好地探索、选择和融合跨模态跨层互补。实验表明,所提出的 CNN 框架在其他最先进的方法上有显著且一致的改进。

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验