• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于联合软-硬注意力的自监督单目深度估计。

Joint Soft-Hard Attention for Self-Supervised Monocular Depth Estimation.

机构信息

University of Chinese Academy of Sciences, Beijing 100049, China.

Shenyang Institute of Computing Technology, Chinese Academy of Sciences, Shenyang 110168, China.

出版信息

Sensors (Basel). 2021 Oct 20;21(21):6956. doi: 10.3390/s21216956.

DOI:10.3390/s21216956
PMID:34770263
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8588100/
Abstract

In recent years, self-supervised monocular depth estimation has gained popularity among researchers because it uses only a single camera at a much lower cost than the direct use of laser sensors to acquire depth. Although monocular self-supervised methods can obtain dense depths, the estimation accuracy needs to be further improved for better applications in scenarios such as autonomous driving and robot perception. In this paper, we innovatively combine soft attention and hard attention with two new ideas to improve self-supervised monocular depth estimation: (1) a soft attention module and (2) a hard attention strategy. We integrate the soft attention module in the model architecture to enhance feature extraction in both spatial and channel dimensions, adding only a small number of parameters. Unlike traditional fusion approaches, we use the hard attention strategy to enhance the fusion of generated multi-scale depth predictions. Further experiments demonstrate that our method can achieve the best self-supervised performance both on the standard KITTI benchmark and the Make3D dataset.

摘要

近年来,自监督单目深度估计在研究人员中变得流行起来,因为它仅使用单个相机,成本远低于直接使用激光传感器来获取深度。尽管单目自监督方法可以获得密集的深度,但为了在自动驾驶和机器人感知等场景中更好地应用,估计精度需要进一步提高。在本文中,我们创新性地将软注意和硬注意与两个新思想结合起来,以提高自监督单目深度估计:(1)软注意模块和(2)硬注意策略。我们将软注意模块集成到模型架构中,以增强空间和通道维度的特征提取,仅增加少量参数。与传统的融合方法不同,我们使用硬注意策略来增强生成的多尺度深度预测的融合。进一步的实验表明,我们的方法在标准 KITTI 基准和 Make3D 数据集上都能实现最佳的自监督性能。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/afb5/8588100/3eb43d9a3006/sensors-21-06956-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/afb5/8588100/b7f127ac6753/sensors-21-06956-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/afb5/8588100/b3b9fd4dc234/sensors-21-06956-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/afb5/8588100/a556565af770/sensors-21-06956-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/afb5/8588100/1f1d812d47ba/sensors-21-06956-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/afb5/8588100/5b4f93ad1f6d/sensors-21-06956-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/afb5/8588100/56a6c0f9c081/sensors-21-06956-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/afb5/8588100/1ee9fd486513/sensors-21-06956-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/afb5/8588100/3eb43d9a3006/sensors-21-06956-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/afb5/8588100/b7f127ac6753/sensors-21-06956-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/afb5/8588100/b3b9fd4dc234/sensors-21-06956-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/afb5/8588100/a556565af770/sensors-21-06956-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/afb5/8588100/1f1d812d47ba/sensors-21-06956-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/afb5/8588100/5b4f93ad1f6d/sensors-21-06956-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/afb5/8588100/56a6c0f9c081/sensors-21-06956-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/afb5/8588100/1ee9fd486513/sensors-21-06956-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/afb5/8588100/3eb43d9a3006/sensors-21-06956-g008.jpg

相似文献

1
Joint Soft-Hard Attention for Self-Supervised Monocular Depth Estimation.基于联合软-硬注意力的自监督单目深度估计。
Sensors (Basel). 2021 Oct 20;21(21):6956. doi: 10.3390/s21216956.
2
MLDA-Net: Multi-Level Dual Attention-Based Network for Self-Supervised Monocular Depth Estimation.MLDA-Net:用于自监督单目深度估计的基于多级双重注意力的网络。
IEEE Trans Image Process. 2021;30:4691-4705. doi: 10.1109/TIP.2021.3074306. Epub 2021 May 3.
3
Multi-Scale Spatial Attention-Guided Monocular Depth Estimation With Semantic Enhancement.具有语义增强的多尺度空间注意力引导单目深度估计
IEEE Trans Image Process. 2021;30:8811-8822. doi: 10.1109/TIP.2021.3120670. Epub 2021 Oct 27.
4
Self-supervised recurrent depth estimation with attention mechanisms.基于注意力机制的自监督循环深度估计
PeerJ Comput Sci. 2022 Jan 31;8:e865. doi: 10.7717/peerj-cs.865. eCollection 2022.
5
Monocular Depth Estimation via Self-Supervised Self-Distillation.通过自监督自蒸馏进行单目深度估计
Sensors (Basel). 2024 Jun 24;24(13):4090. doi: 10.3390/s24134090.
6
Semi-Supervised Adversarial Monocular Depth Estimation.半监督对抗式单目深度估计
IEEE Trans Pattern Anal Mach Intell. 2020 Oct;42(10):2410-2422. doi: 10.1109/TPAMI.2019.2936024. Epub 2019 Aug 20.
7
SFA-MDEN: Semantic-Feature-Aided Monocular Depth Estimation Network Using Dual Branches.SFA-MDEN:基于语义特征辅助的双通道单目深度估计网络。
Sensors (Basel). 2021 Aug 13;21(16):5476. doi: 10.3390/s21165476.
8
Unsupervised Monocular Depth Estimation With Channel and Spatial Attention.基于通道和空间注意力的无监督单目深度估计
IEEE Trans Neural Netw Learn Syst. 2024 Jun;35(6):7860-7870. doi: 10.1109/TNNLS.2022.3221416. Epub 2024 Jun 3.
9
Online supervised attention-based recurrent depth estimation from monocular video.基于在线监督注意力机制的单目视频递归深度估计
PeerJ Comput Sci. 2020 Nov 23;6:e317. doi: 10.7717/peerj-cs.317. eCollection 2020.
10
PMIndoor: Pose Rectified Network and Multiple Loss Functions for Self-Supervised Monocular Indoor Depth Estimation.PMIndoor:用于自监督单目室内深度估计的姿态校正网络和多重损失函数
Sensors (Basel). 2023 Oct 30;23(21):8821. doi: 10.3390/s23218821.

引用本文的文献

1
Film and Video Quality Optimization Using Attention Mechanism-Embedded Lightweight Neural Network Model.利用注意力机制嵌入的轻量级神经网络模型优化电影和视频质量。
Comput Intell Neurosci. 2022 Jun 8;2022:8229580. doi: 10.1155/2022/8229580. eCollection 2022.

本文引用的文献

1
Every Pixel Counts ++: Joint Learning of Geometry and Motion with 3D Holistic Understanding.每一个像素都很重要++:通过3D整体理解进行几何与运动的联合学习。
IEEE Trans Pattern Anal Mach Intell. 2019 Jul 23. doi: 10.1109/TPAMI.2019.2930258.
2
Deep Ordinal Regression Network for Monocular Depth Estimation.用于单目深度估计的深度序数回归网络
Proc IEEE Comput Soc Conf Comput Vis Pattern Recognit. 2018 Jun;2018:2002-2011. doi: 10.1109/CVPR.2018.00214. Epub 2018 Dec 17.
3
Learning Depth from Single Monocular Images Using Deep Convolutional Neural Fields.
利用深度卷积神经场从单目图像中学习深度。
IEEE Trans Pattern Anal Mach Intell. 2016 Oct;38(10):2024-39. doi: 10.1109/TPAMI.2015.2505283. Epub 2015 Dec 3.
4
Depth Transfer: Depth Extraction from Video Using Non-Parametric Sampling.深度迁移:使用非参数采样从视频中提取深度。
IEEE Trans Pattern Anal Mach Intell. 2014 Nov;36(11):2144-58. doi: 10.1109/TPAMI.2014.2316835.
5
Make3D: learning 3D scene structure from a single still image.Make3D:从单张静止图像学习3D场景结构。
IEEE Trans Pattern Anal Mach Intell. 2009 May;31(5):824-40. doi: 10.1109/TPAMI.2008.132.