• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于通道和空间注意力的无监督单目深度估计

Unsupervised Monocular Depth Estimation With Channel and Spatial Attention.

作者信息

Wang Zhuping, Dai Xinke, Guo Zhanyu, Huang Chao, Zhang Hao

出版信息

IEEE Trans Neural Netw Learn Syst. 2024 Jun;35(6):7860-7870. doi: 10.1109/TNNLS.2022.3221416. Epub 2024 Jun 3.

DOI:10.1109/TNNLS.2022.3221416
PMID:36459609
Abstract

Understanding 3-D scene geometry from videos is a fundamental topic in visual perception. In this article, we propose an unsupervised monocular depth and camera motion estimation framework using unlabeled monocular videos to overcome the limitation of acquiring per-pixel ground-truth depth at scale. The photometric loss couples the depth network and pose network together and is essential to the unsupervised method, which is based on warping nearby views to target using the estimated depth and pose. We introduce the channelwise attention mechanism to dig into the relationship between channels and introduce the spatialwise attention mechanism to utilize the inner-spatial relationship of features. Both of them applied in depth networks can better activate the feature information between different convolutional layers and extract more discriminative features. In addition, we apply the Sobel boundary to our edge-aware smoothness for more reasonable accuracy, and clearer boundaries and structures. All of these help to close the gap with fully supervised methods and show high-quality state-of-the-art results on the KITTI benchmark and great generalization performance on the Make3D dataset.

摘要

从视频中理解三维场景几何是视觉感知中的一个基本课题。在本文中,我们提出了一个无监督的单目深度和相机运动估计框架,该框架使用未标记的单目视频,以克服大规模获取逐像素地面真值深度的局限性。光度损失将深度网络和姿态网络耦合在一起,对于基于使用估计的深度和姿态将附近视图扭曲到目标的无监督方法至关重要。我们引入通道注意力机制来深入研究通道之间的关系,并引入空间注意力机制来利用特征的内部空间关系。将它们两者应用于深度网络可以更好地激活不同卷积层之间的特征信息,并提取更具判别力的特征。此外,我们将Sobel边界应用于边缘感知平滑度,以获得更合理的精度、更清晰的边界和结构。所有这些都有助于缩小与完全监督方法的差距,并在KITTI基准测试中展示高质量的最新结果,以及在Make3D数据集上具有出色的泛化性能。

相似文献

1
Unsupervised Monocular Depth Estimation With Channel and Spatial Attention.基于通道和空间注意力的无监督单目深度估计
IEEE Trans Neural Netw Learn Syst. 2024 Jun;35(6):7860-7870. doi: 10.1109/TNNLS.2022.3221416. Epub 2024 Jun 3.
2
Joint Unsupervised Learning of Depth, Pose, Ground Normal Vector and Ground Segmentation by a Monocular Camera Sensor.基于单目相机传感器的深度、姿势、地面法向量和地面分割的联合无监督学习。
Sensors (Basel). 2020 Jul 3;20(13):3737. doi: 10.3390/s20133737.
3
Multi-Scale Spatial Attention-Guided Monocular Depth Estimation With Semantic Enhancement.具有语义增强的多尺度空间注意力引导单目深度估计
IEEE Trans Image Process. 2021;30:8811-8822. doi: 10.1109/TIP.2021.3120670. Epub 2021 Oct 27.
4
An Unsupervised Monocular Visual Odometry Based on Multi-Scale Modeling.一种基于多尺度建模的无监督单目视觉里程计
Sensors (Basel). 2022 Jul 11;22(14):5193. doi: 10.3390/s22145193.
5
RAUM-VO: Rotational Adjusted Unsupervised Monocular Visual Odometry.RAUM-VO:旋转调整无监督单目视觉里程计。
Sensors (Basel). 2022 Mar 30;22(7):2651. doi: 10.3390/s22072651.
6
Joint Soft-Hard Attention for Self-Supervised Monocular Depth Estimation.基于联合软-硬注意力的自监督单目深度估计。
Sensors (Basel). 2021 Oct 20;21(21):6956. doi: 10.3390/s21216956.
7
SFA-MDEN: Semantic-Feature-Aided Monocular Depth Estimation Network Using Dual Branches.SFA-MDEN:基于语义特征辅助的双通道单目深度估计网络。
Sensors (Basel). 2021 Aug 13;21(16):5476. doi: 10.3390/s21165476.
8
EndoSLAM dataset and an unsupervised monocular visual odometry and depth estimation approach for endoscopic videos.内镜 SLAM 数据集和一种用于内镜视频的无监督单目视觉里程计和深度估计方法。
Med Image Anal. 2021 Jul;71:102058. doi: 10.1016/j.media.2021.102058. Epub 2021 Apr 15.
9
SENSE: Self-Evolving Learning for Self-Supervised Monocular Depth Estimation.SENSE:用于自监督单目深度估计的自进化学习
IEEE Trans Image Process. 2024;33:439-450. doi: 10.1109/TIP.2023.3338053. Epub 2023 Dec 29.
10
Unsupervised Monocular Depth Estimation from Light Field Image.基于光场图像的无监督单目深度估计
IEEE Trans Image Process. 2019 Oct 3. doi: 10.1109/TIP.2019.2944343.

引用本文的文献

1
Supervised deep learning-based paradigm to screen the enhanced oil recovery scenarios.基于监督深度学习的方法筛选强化采油场景。
Sci Rep. 2023 Mar 25;13(1):4892. doi: 10.1038/s41598-023-32187-2.