• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于注意力的分层多模态融合实现高分辨率深度图成像

High-Resolution Depth Maps Imaging via Attention-Based Hierarchical Multi-Modal Fusion.

作者信息

Zhong Zhiwei, Liu Xianming, Jiang Junjun, Zhao Debin, Chen Zhiwen, Ji Xiangyang

出版信息

IEEE Trans Image Process. 2022;31:648-663. doi: 10.1109/TIP.2021.3131041. Epub 2021 Dec 28.

DOI:10.1109/TIP.2021.3131041
PMID:34878976
Abstract

Depth map records distance between the viewpoint and objects in the scene, which plays a critical role in many real-world applications. However, depth map captured by consumer-grade RGB-D cameras suffers from low spatial resolution. Guided depth map super-resolution (DSR) is a popular approach to address this problem, which attempts to restore a high-resolution (HR) depth map from the input low-resolution (LR) depth and its coupled HR RGB image that serves as the guidance. The most challenging issue for guided DSR is how to correctly select consistent structures and propagate them, and properly handle inconsistent ones. In this paper, we propose a novel attention-based hierarchical multi-modal fusion (AHMF) network for guided DSR. Specifically, to effectively extract and combine relevant information from LR depth and HR guidance, we propose a multi-modal attention based fusion (MMAF) strategy for hierarchical convolutional layers, including a feature enhancement block to select valuable features and a feature recalibration block to unify the similarity metrics of modalities with different appearance characteristics. Furthermore, we propose a bi-directional hierarchical feature collaboration (BHFC) module to fully leverage low-level spatial information and high-level structure information among multi-scale features. Experimental results show that our approach outperforms state-of-the-art methods in terms of reconstruction accuracy, running speed and memory efficiency.

摘要

深度图记录场景中视点与物体之间的距离,这在许多实际应用中起着关键作用。然而,消费级RGB-D相机捕获的深度图存在空间分辨率低的问题。引导式深度图超分辨率(DSR)是解决这一问题的一种流行方法,它试图从输入的低分辨率(LR)深度图及其作为引导的耦合高分辨率(HR)RGB图像中恢复高分辨率(HR)深度图。引导式DSR最具挑战性的问题是如何正确选择一致的结构并进行传播,以及如何妥善处理不一致的结构。在本文中,我们提出了一种用于引导式DSR的基于注意力的新型分层多模态融合(AHMF)网络。具体而言,为了有效地从LR深度图和HR引导中提取并组合相关信息,我们针对分层卷积层提出了一种基于多模态注意力的融合(MMAF)策略,包括一个用于选择有价值特征的特征增强块和一个用于统一具有不同外观特征的模态相似性度量的特征重新校准块。此外,我们提出了一个双向分层特征协作(BHFC)模块,以充分利用多尺度特征之间的低级空间信息和高级结构信息。实验结果表明,我们的方法在重建精度、运行速度和内存效率方面优于现有方法。

相似文献

1
High-Resolution Depth Maps Imaging via Attention-Based Hierarchical Multi-Modal Fusion.基于注意力的分层多模态融合实现高分辨率深度图成像
IEEE Trans Image Process. 2022;31:648-663. doi: 10.1109/TIP.2021.3131041. Epub 2021 Dec 28.
2
Hierarchical Features Driven Residual Learning for Depth Map Super-Resolution.用于深度图超分辨率的层次特征驱动残差学习
IEEE Trans Image Process. 2018 Dec 17. doi: 10.1109/TIP.2018.2887029.
3
Fully Cross-Attention Transformer for Guided Depth Super-Resolution.基于全交叉注意力变换的导向深度超分辨率
Sensors (Basel). 2023 Mar 2;23(5):2723. doi: 10.3390/s23052723.
4
SLMSF-Net: A Semantic Localization and Multi-Scale Fusion Network for RGB-D Salient Object Detection.SLMSF-Net:用于RGB-D显著目标检测的语义定位与多尺度融合网络
Sensors (Basel). 2024 Feb 8;24(4):1117. doi: 10.3390/s24041117.
5
Hierarchical Alternate Interaction Network for RGB-D Salient Object Detection.用于RGB-D显著目标检测的分层交替交互网络
IEEE Trans Image Process. 2021;30:3528-3542. doi: 10.1109/TIP.2021.3062689. Epub 2021 Mar 11.
6
DMRA: Depth-Induced Multi-Scale Recurrent Attention Network for RGB-D Saliency Detection.DMRA:用于 RGB-D 显著度检测的深度诱导多尺度递归注意网络。
IEEE Trans Image Process. 2022;31:2321-2336. doi: 10.1109/TIP.2022.3154931. Epub 2022 Mar 11.
7
RGB-T Salient Object Detection via Fusing Multi-level CNN Features.基于融合多级卷积神经网络特征的RGB-T显著目标检测
IEEE Trans Image Process. 2019 Dec 17. doi: 10.1109/TIP.2019.2959253.
8
Color-Guided Depth Map Super-Resolution Using a Dual-Branch Multi-Scale Residual Network with Channel Interaction.基于双通道多尺度残差网络和通道交互的彩色引导深度图超分辨率方法。
Sensors (Basel). 2020 Mar 11;20(6):1560. doi: 10.3390/s20061560.
9
Deep Convolutional Neural Network for Multi-Modal Image Restoration and Fusion.深度卷积神经网络的多模态图像恢复与融合。
IEEE Trans Pattern Anal Mach Intell. 2021 Oct;43(10):3333-3348. doi: 10.1109/TPAMI.2020.2984244. Epub 2021 Sep 2.
10
CrossFuNet: RGB and Depth Cross-Fusion Network for Hand Pose Estimation.CrossFuNet:用于手部姿势估计的 RGB 和深度交叉融合网络。
Sensors (Basel). 2021 Sep 11;21(18):6095. doi: 10.3390/s21186095.

引用本文的文献

1
IGAF: Incremental Guided Attention Fusion for Depth Super-Resolution.IGAF:用于深度超分辨率的增量引导注意力融合
Sensors (Basel). 2024 Dec 24;25(1):24. doi: 10.3390/s25010024.
2
Multi-scale Fusion of Stretched Infrared and Visible Images.拉伸红外与可见光图像的多尺度融合。
Sensors (Basel). 2022 Sep 2;22(17):6660. doi: 10.3390/s22176660.
3
A Generic Framework for Depth Reconstruction Enhancement.深度重建增强的通用框架。
J Imaging. 2022 May 16;8(5):138. doi: 10.3390/jimaging8050138.