• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

IGAF:用于深度超分辨率的增量引导注意力融合

IGAF: Incremental Guided Attention Fusion for Depth Super-Resolution.

作者信息

Tragakis Athanasios, Kaul Chaitanya, Mitchell Kevin J, Dai Hang, Murray-Smith Roderick, Faccio Daniele

机构信息

School of Physics and Astronomy, University of Glasgow, Glasgow G12 8QQ, UK.

School of Computing Science, University of Glasgow, Glasgow G12 8QQ, UK.

出版信息

Sensors (Basel). 2024 Dec 24;25(1):24. doi: 10.3390/s25010024.

DOI:10.3390/s25010024
PMID:39796813
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11723017/
Abstract

Accurate depth estimation is crucial for many fields, including robotics, navigation, and medical imaging. However, conventional depth sensors often produce low-resolution (LR) depth maps, making detailed scene perception challenging. To address this, enhancing LR depth maps to high-resolution (HR) ones has become essential, guided by HR-structured inputs like RGB or grayscale images. We propose a novel sensor fusion methodology for guided depth super-resolution (GDSR), a technique that combines LR depth maps with HR images to estimate detailed HR depth maps. Our key contribution is the Incremental guided attention fusion (IGAF) module, which effectively learns to fuse features from RGB images and LR depth maps, producing accurate HR depth maps. Using IGAF, we build a robust super-resolution model and evaluate it on multiple benchmark datasets. Our model achieves state-of-the-art results compared to all baseline models on the NYU v2 dataset for ×4, ×8, and ×16 upsampling. It also outperforms all baselines in a zero-shot setting on the Middlebury, Lu, and RGB-D-D datasets. Code, environments, and models are available on GitHub.

摘要

精确的深度估计对于包括机器人技术、导航和医学成像在内的许多领域至关重要。然而,传统的深度传感器通常会生成低分辨率(LR)深度图,这使得详细的场景感知具有挑战性。为了解决这个问题,在RGB或灰度图像等高分辨率(HR)结构输入的引导下,将低分辨率深度图增强为高分辨率深度图变得至关重要。我们提出了一种用于引导深度超分辨率(GDSR)的新型传感器融合方法,该技术将低分辨率深度图与高分辨率图像相结合,以估计详细的高分辨率深度图。我们的关键贡献是增量引导注意力融合(IGAF)模块,它有效地学习融合来自RGB图像和低分辨率深度图的特征,从而生成准确的高分辨率深度图。使用IGAF,我们构建了一个强大的超分辨率模型,并在多个基准数据集上对其进行评估。在纽约大学v2数据集上进行×4、×8和×16上采样时,我们的模型与所有基线模型相比取得了领先的结果。在米德尔伯里、卢和RGB-D-D数据集的零样本设置中,它也优于所有基线。代码、环境和模型可在GitHub上获取。

相似文献

1
IGAF: Incremental Guided Attention Fusion for Depth Super-Resolution.IGAF:用于深度超分辨率的增量引导注意力融合
Sensors (Basel). 2024 Dec 24;25(1):24. doi: 10.3390/s25010024.
2
High-Resolution Depth Maps Imaging via Attention-Based Hierarchical Multi-Modal Fusion.基于注意力的分层多模态融合实现高分辨率深度图成像
IEEE Trans Image Process. 2022;31:648-663. doi: 10.1109/TIP.2021.3131041. Epub 2021 Dec 28.
3
Fully Cross-Attention Transformer for Guided Depth Super-Resolution.基于全交叉注意力变换的导向深度超分辨率
Sensors (Basel). 2023 Mar 2;23(5):2723. doi: 10.3390/s23052723.
4
CDNet: Complementary Depth Network for RGB-D Salient Object Detection.CDNet:用于RGB-D显著目标检测的互补深度网络。
IEEE Trans Image Process. 2021;30:3376-3390. doi: 10.1109/TIP.2021.3060167. Epub 2021 Mar 9.
5
Moving Object Detection Based on Fusion of Depth Information and RGB Features.基于深度信息与 RGB 特征融合的运动目标检测。
Sensors (Basel). 2022 Jun 22;22(13):4702. doi: 10.3390/s22134702.
6
Hyperspectral Image Super Resolution With Real Unaligned RGB Guidance.基于真实未对齐RGB引导的高光谱图像超分辨率
IEEE Trans Neural Netw Learn Syst. 2025 Feb;36(2):2999-3011. doi: 10.1109/TNNLS.2023.3340561. Epub 2025 Feb 6.
7
Deep Unsupervised Fusion Learning for Hyperspectral Image Super Resolution.用于高光谱图像超分辨率的深度无监督融合学习
Sensors (Basel). 2021 Mar 28;21(7):2348. doi: 10.3390/s21072348.
8
Method of depth simulation imaging and depth image super-resolution reconstruction for a 2D/3D compatible CMOS image sensor.用于二维/三维兼容互补金属氧化物半导体图像传感器的深度模拟成像和深度图像超分辨率重建方法。
Appl Opt. 2023 Jun 10;62(17):4439-4454. doi: 10.1364/AO.493280.
9
SGSR: style-subnets-assisted generative latent bank for large-factor super-resolution with registered medical image dataset.SGSR:基于风格子网的生成式潜在银行,用于具有注册医学图像数据集的大因子超分辨率。
Int J Comput Assist Radiol Surg. 2024 Mar;19(3):493-506. doi: 10.1007/s11548-023-03037-3. Epub 2023 Dec 21.
10
Multimodal super-resolved q-space deep learning.多模态超分辨率 q 空间深度学习。
Med Image Anal. 2021 Jul;71:102085. doi: 10.1016/j.media.2021.102085. Epub 2021 Apr 21.

本文引用的文献

1
Multiscale Attention Fusion for Depth Map Super-Resolution Generative Adversarial Networks.用于深度图超分辨率生成对抗网络的多尺度注意力融合
Entropy (Basel). 2023 May 23;25(6):836. doi: 10.3390/e25060836.
2
High-Resolution Depth Maps Imaging via Attention-Based Hierarchical Multi-Modal Fusion.基于注意力的分层多模态融合实现高分辨率深度图成像
IEEE Trans Image Process. 2022;31:648-663. doi: 10.1109/TIP.2021.3131041. Epub 2021 Dec 28.
3
Deep Convolutional Neural Network for Multi-Modal Image Restoration and Fusion.深度卷积神经网络的多模态图像恢复与融合。
IEEE Trans Pattern Anal Mach Intell. 2021 Oct;43(10):3333-3348. doi: 10.1109/TPAMI.2020.2984244. Epub 2021 Sep 2.
4
Weighted Guided Image Filtering with Steering Kernel.带导向核的加权引导图像滤波
IEEE Trans Image Process. 2019 Jul 19. doi: 10.1109/TIP.2019.2928631.
5
Joint Image Filtering with Deep Convolutional Networks.基于深度卷积网络的联合图像滤波
IEEE Trans Pattern Anal Mach Intell. 2019 Aug;41(8):1909-1923. doi: 10.1109/TPAMI.2018.2890623. Epub 2019 Jan 1.
6
Edge-Preserving Depth Map Upsampling by Joint Trilateral Filter.基于联合三边滤波的边缘保持深度图上采样。
IEEE Trans Cybern. 2018 Jan;48(1):371-384. doi: 10.1109/TCYB.2016.2637661. Epub 2017 Jan 24.
7
High-quality depth map upsampling and completion for RGB-D cameras.高质量深度图的 RGB-D 相机上采样和补全。
IEEE Trans Image Process. 2014 Dec;23(12):5559-72. doi: 10.1109/TIP.2014.2361034.
8
A fast and fully automatic method for cerebrovascular segmentation on time-of-flight (TOF) MRA image.一种快速全自动的基于时间飞跃(TOF)MRA 图像的脑血管分割方法。
J Digit Imaging. 2011 Aug;24(4):609-25. doi: 10.1007/s10278-010-9326-1.
9
Time-of-Flight 3-D endoscopy.飞行时间三维内镜检查
Med Image Comput Comput Assist Interv. 2009;12(Pt 1):467-74. doi: 10.1007/978-3-642-04268-3_58.