• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

单目深度估计:轻量级卷积和矩阵胶囊特征融合网络。

Monocular Depth Estimation: Lightweight Convolutional and Matrix Capsule Feature-Fusion Network.

机构信息

College of Information Science & Technology, Beijing University of Chemical Technology, Beijing 100029, China.

出版信息

Sensors (Basel). 2022 Aug 23;22(17):6344. doi: 10.3390/s22176344.

DOI:10.3390/s22176344
PMID:36080801
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9459913/
Abstract

This paper reports a study that aims to solve the problem of the weak adaptability to angle transformation of current monocular depth estimation algorithms. These algorithms are based on convolutional neural networks (CNNs) but produce results lacking in estimation accuracy and robustness. The paper proposes a lightweight network based on convolution and capsule feature fusion (CNNapsule). First, the paper introduces a fusion block module that integrates CNN features and matrix capsule features to improve the adaptability of the network to perspective transformations. The fusion and deconvolution features are fused through skip connections to generate a depth image. In addition, the corresponding loss function is designed according to the long-tail distribution, gradient similarity, and structural similarity of the datasets. Finally, the results are compared with the methods applied to the NYU Depth V2 and KITTI datasets and show that our proposed method has better accuracy on the C1 and C2 indices and a better visual effect than traditional methods and deep learning methods without transfer learning. The number of trainable parameters required by this method is 65% lower than that required by methods presented in the literature. The generalization of this method is verified via the comparative testing of the data collected from the internet and mobile phones.

摘要

本文报道了一项研究,旨在解决当前单目深度估计算法对角度变换的弱适应性问题。这些算法基于卷积神经网络(CNN),但产生的结果在估计精度和鲁棒性方面存在不足。本文提出了一种基于卷积和胶囊特征融合的轻量级网络(CNNapsule)。首先,本文引入了一个融合块模块,该模块集成了 CNN 特征和矩阵胶囊特征,以提高网络对透视变换的适应性。融合和反卷积特征通过跳过连接进行融合,以生成深度图像。此外,根据数据集的长尾分布、梯度相似性和结构相似性设计了相应的损失函数。最后,将结果与应用于 NYU Depth V2 和 KITTI 数据集的方法进行比较,结果表明,与传统方法和不进行迁移学习的深度学习方法相比,我们提出的方法在 C1 和 C2 指标上具有更好的准确性和更好的视觉效果。该方法所需的可训练参数数量比文献中提出的方法低 65%。通过对从互联网和手机收集的数据进行比较测试,验证了该方法的泛化能力。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3b5e/9459913/e3996ae77ab3/sensors-22-06344-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3b5e/9459913/19b0d7a7c9e6/sensors-22-06344-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3b5e/9459913/c5097a922b9b/sensors-22-06344-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3b5e/9459913/f82a77d5e8ee/sensors-22-06344-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3b5e/9459913/2753b2714d42/sensors-22-06344-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3b5e/9459913/d9a5c9f0bec3/sensors-22-06344-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3b5e/9459913/6ab660e8aeca/sensors-22-06344-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3b5e/9459913/fe68fc842f0b/sensors-22-06344-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3b5e/9459913/9548d1b378bc/sensors-22-06344-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3b5e/9459913/ccdd9131a935/sensors-22-06344-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3b5e/9459913/e3996ae77ab3/sensors-22-06344-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3b5e/9459913/19b0d7a7c9e6/sensors-22-06344-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3b5e/9459913/c5097a922b9b/sensors-22-06344-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3b5e/9459913/f82a77d5e8ee/sensors-22-06344-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3b5e/9459913/2753b2714d42/sensors-22-06344-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3b5e/9459913/d9a5c9f0bec3/sensors-22-06344-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3b5e/9459913/6ab660e8aeca/sensors-22-06344-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3b5e/9459913/fe68fc842f0b/sensors-22-06344-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3b5e/9459913/9548d1b378bc/sensors-22-06344-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3b5e/9459913/ccdd9131a935/sensors-22-06344-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3b5e/9459913/e3996ae77ab3/sensors-22-06344-g010.jpg

相似文献

1
Monocular Depth Estimation: Lightweight Convolutional and Matrix Capsule Feature-Fusion Network.单目深度估计:轻量级卷积和矩阵胶囊特征融合网络。
Sensors (Basel). 2022 Aug 23;22(17):6344. doi: 10.3390/s22176344.
2
Lightweight monocular depth estimation using a fusion-improved transformer.使用融合改进型变压器的轻量级单目深度估计
Sci Rep. 2024 Sep 28;14(1):22472. doi: 10.1038/s41598-024-72682-8.
3
Depth Estimation from Light Field Geometry Using Convolutional Neural Networks.基于卷积神经网络的光场几何深度估计
Sensors (Basel). 2021 Sep 10;21(18):6061. doi: 10.3390/s21186061.
4
LightMixer: A novel lightweight convolutional neural network for tomato disease detection.LightMixer:一种用于番茄病害检测的新型轻量级卷积神经网络。
Front Plant Sci. 2023 May 9;14:1166296. doi: 10.3389/fpls.2023.1166296. eCollection 2023.
5
Deep Ordinal Regression Network for Monocular Depth Estimation.用于单目深度估计的深度序数回归网络
Proc IEEE Comput Soc Conf Comput Vis Pattern Recognit. 2018 Jun;2018:2002-2011. doi: 10.1109/CVPR.2018.00214. Epub 2018 Dec 17.
6
AMENet is a monocular depth estimation network designed for automatic stereoscopic display.AMENet是一种为自动立体显示而设计的单目深度估计网络。
Sci Rep. 2024 Mar 11;14(1):5868. doi: 10.1038/s41598-024-56095-1.
7
Monocular Depth Estimation Using a Laplacian Image Pyramid with Local Planar Guidance Layers.基于拉普拉斯图像金字塔和局部平面引导层的单目深度估计
Sensors (Basel). 2023 Jan 11;23(2):845. doi: 10.3390/s23020845.
8
Monocular Depth Estimation via Self-Supervised Self-Distillation.通过自监督自蒸馏进行单目深度估计
Sensors (Basel). 2024 Jun 24;24(13):4090. doi: 10.3390/s24134090.
9
RT-ViT: Real-Time Monocular Depth Estimation Using Lightweight Vision Transformers.RT-ViT:基于轻量级视觉Transformer 的实时单目深度估计。
Sensors (Basel). 2022 May 19;22(10):3849. doi: 10.3390/s22103849.
10
SFA-MDEN: Semantic-Feature-Aided Monocular Depth Estimation Network Using Dual Branches.SFA-MDEN:基于语义特征辅助的双通道单目深度估计网络。
Sensors (Basel). 2021 Aug 13;21(16):5476. doi: 10.3390/s21165476.

引用本文的文献

1
An Adaptive Refinement Scheme for Depth Estimation Networks.自适应细化网络的深度估计。
Sensors (Basel). 2022 Dec 13;22(24):9755. doi: 10.3390/s22249755.

本文引用的文献

1
Mobile3DRecon: Real-time Monocular 3D Reconstruction on a Mobile Phone.移动3D重建:手机上的实时单目3D重建
IEEE Trans Vis Comput Graph. 2020 Dec;26(12):3446-3456. doi: 10.1109/TVCG.2020.3023634. Epub 2020 Nov 10.
2
Superb Monocular Depth Estimation Based on Transfer Learning and Surface Normal Guidance.基于迁移学习和表面法向导引的卓越单目深度估计。
Sensors (Basel). 2020 Aug 27;20(17):4856. doi: 10.3390/s20174856.
3
Semi-Supervised Adversarial Monocular Depth Estimation.半监督对抗式单目深度估计
IEEE Trans Pattern Anal Mach Intell. 2020 Oct;42(10):2410-2422. doi: 10.1109/TPAMI.2019.2936024. Epub 2019 Aug 20.
4
Deep Ordinal Regression Network for Monocular Depth Estimation.用于单目深度估计的深度序数回归网络
Proc IEEE Comput Soc Conf Comput Vis Pattern Recognit. 2018 Jun;2018:2002-2011. doi: 10.1109/CVPR.2018.00214. Epub 2018 Dec 17.
5
Depth Estimation and Semantic Segmentation from a Single RGB Image Using a Hybrid Convolutional Neural Network.使用混合卷积神经网络从单张RGB图像进行深度估计和语义分割
Sensors (Basel). 2019 Apr 15;19(8):1795. doi: 10.3390/s19081795.