• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于拉普拉斯图像金字塔和局部平面引导层的单目深度估计

Monocular Depth Estimation Using a Laplacian Image Pyramid with Local Planar Guidance Layers.

机构信息

Department of Smart Car Engineering, Chungbuk National University, 1 Chungdae-ro, Seowon-gu, Cheongju-si 28644, Republic of Korea.

Department of Intelligent Systems and Robotics, Chungbuk National University, 1 Chungdae-ro, Seowon-gu, Cheongju-si 28644, Republic of Korea.

出版信息

Sensors (Basel). 2023 Jan 11;23(2):845. doi: 10.3390/s23020845.

DOI:10.3390/s23020845
PMID:36679643
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9862419/
Abstract

It is important to estimate the exact depth from 2D images, and many studies have been conducted for a long period of time to solve depth estimation problems. Recently, as research on estimating depth from monocular camera images based on deep learning is progressing, research for estimating accurate depths using various techniques is being conducted. However, depth estimation from 2D images has been a problem in predicting the boundary between objects. In this paper, we aim to predict sophisticated depths by emphasizing the precise boundaries between objects. We propose a depth estimation network with encoder-decoder structures using the Laplacian pyramid and local planar guidance method. In the process of upsampling the learned features using the encoder, the purpose of this step is to obtain a clearer depth map by guiding a more sophisticated boundary of an object using the Laplacian pyramid and local planar guidance techniques. We train and test our models with KITTI and NYU Depth V2 datasets. The proposed network constructs a DNN using only convolution and uses the ConvNext networks as a backbone. A trained model shows the performance of the absolute relative error (Abs_rel) 0.054 and root mean square error (RMSE) 2.252 based on the KITTI dataset and absolute relative error (Abs_rel) 0.102 and root mean square error 0.355 based on the NYU Depth V2 dataset. On the state-of-the-art monocular depth estimation, our network performance shows the fifth-best performance based on the KITTI Eigen split and the eighth-best performance based on the NYU Depth V2.

摘要

从二维图像估计准确的深度很重要,长期以来,许多研究都致力于解决深度估计问题。最近,随着基于深度学习的单目相机图像深度估计研究的进展,使用各种技术进行准确深度估计的研究正在进行。然而,从二维图像进行深度估计一直是预测物体边界的问题。在本文中,我们旨在通过强调物体之间的精确边界来预测复杂的深度。我们提出了一种具有拉普拉斯金字塔和局部平面引导方法的编解码器结构的深度估计网络。在使用编码器对学习到的特征进行上采样的过程中,该步骤的目的是通过使用拉普拉斯金字塔和局部平面引导技术引导更复杂的物体边界,从而获得更清晰的深度图。我们使用 KITTI 和 NYU Depth V2 数据集对我们的模型进行训练和测试。该网络仅使用卷积构建 DNN,并使用 ConvNext 网络作为骨干。训练好的模型在 KITTI 数据集上的绝对相对误差 (Abs_rel) 为 0.054,均方根误差 (RMSE) 为 2.252,在 NYU Depth V2 数据集上的绝对相对误差 (Abs_rel) 为 0.102,均方根误差为 0.355。在单目深度估计的最新技术中,我们的网络性能在 KITTI Eigen 拆分中排名第五,在 NYU Depth V2 中排名第八。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b867/9862419/a9f9953ec217/sensors-23-00845-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b867/9862419/a542411c79ee/sensors-23-00845-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b867/9862419/40903f121e03/sensors-23-00845-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b867/9862419/7daefa8b97fd/sensors-23-00845-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b867/9862419/e538602b12cb/sensors-23-00845-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b867/9862419/2ce69af26459/sensors-23-00845-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b867/9862419/b36e651ed35b/sensors-23-00845-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b867/9862419/49ca46c34664/sensors-23-00845-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b867/9862419/a9f9953ec217/sensors-23-00845-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b867/9862419/a542411c79ee/sensors-23-00845-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b867/9862419/40903f121e03/sensors-23-00845-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b867/9862419/7daefa8b97fd/sensors-23-00845-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b867/9862419/e538602b12cb/sensors-23-00845-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b867/9862419/2ce69af26459/sensors-23-00845-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b867/9862419/b36e651ed35b/sensors-23-00845-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b867/9862419/49ca46c34664/sensors-23-00845-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/b867/9862419/a9f9953ec217/sensors-23-00845-g008.jpg

相似文献

1
Monocular Depth Estimation Using a Laplacian Image Pyramid with Local Planar Guidance Layers.基于拉普拉斯图像金字塔和局部平面引导层的单目深度估计
Sensors (Basel). 2023 Jan 11;23(2):845. doi: 10.3390/s23020845.
2
LapUNet: a novel approach to monocular depth estimation using dynamic laplacian residual U-shape networks.LapUNet:一种使用动态拉普拉斯残差U型网络进行单目深度估计的新方法。
Sci Rep. 2024 Oct 9;14(1):23544. doi: 10.1038/s41598-024-74445-x.
3
Laplacian Pyramid Neural Network for Dense Continuous-Value Regression for Complex Scenes.用于复杂场景密集连续值回归的拉普拉斯金字塔神经网络。
IEEE Trans Neural Netw Learn Syst. 2021 Nov;32(11):5034-5046. doi: 10.1109/TNNLS.2020.3026669. Epub 2021 Oct 27.
4
Deep Monocular Depth Estimation Based on Content and Contextual Features.基于内容和上下文特征的深度单目深度估计。
Sensors (Basel). 2023 Mar 8;23(6):2919. doi: 10.3390/s23062919.
5
Dense monocular depth estimation for stereoscopic vision based on pyramid transformer and multi-scale feature fusion.基于金字塔变换器和多尺度特征融合的立体视觉密集单目深度估计
Sci Rep. 2024 Mar 25;14(1):7037. doi: 10.1038/s41598-024-57908-z.
6
A Novel Method for Monocular Depth Estimation Using an Hourglass Neck Module.一种使用沙漏颈部模块进行单目深度估计的新方法。
Sensors (Basel). 2024 Feb 18;24(4):1312. doi: 10.3390/s24041312.
7
RT-ViT: Real-Time Monocular Depth Estimation Using Lightweight Vision Transformers.RT-ViT:基于轻量级视觉Transformer 的实时单目深度估计。
Sensors (Basel). 2022 May 19;22(10):3849. doi: 10.3390/s22103849.
8
GFI-Net: Global Feature Interaction Network for Monocular Depth Estimation.GFI-Net:用于单目深度估计的全局特征交互网络。
Entropy (Basel). 2023 Feb 26;25(3):421. doi: 10.3390/e25030421.
9
DCPNet: A Densely Connected Pyramid Network for Monocular Depth Estimation.DCPNet:用于单目深度估计的密集连接金字塔网络。
Sensors (Basel). 2021 Oct 13;21(20):6780. doi: 10.3390/s21206780.
10
Monocular Depth Estimation via Self-Supervised Self-Distillation.通过自监督自蒸馏进行单目深度估计
Sensors (Basel). 2024 Jun 24;24(13):4090. doi: 10.3390/s24134090.

本文引用的文献

1
BinsFormer: Revisiting Adaptive Bins for Monocular Depth Estimation.BinsFormer:重新审视用于单目深度估计的自适应 bins
IEEE Trans Image Process. 2024;33:3964-3976. doi: 10.1109/TIP.2024.3416065. Epub 2024 Jun 28.
2
Deep Ordinal Regression Network for Monocular Depth Estimation.用于单目深度估计的深度序数回归网络
Proc IEEE Comput Soc Conf Comput Vis Pattern Recognit. 2018 Jun;2018:2002-2011. doi: 10.1109/CVPR.2018.00214. Epub 2018 Dec 17.
3
DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs.
DeepLab:基于深度卷积网络、空洞卷积和全连接条件随机场的语义图像分割。
IEEE Trans Pattern Anal Mach Intell. 2018 Apr;40(4):834-848. doi: 10.1109/TPAMI.2017.2699184. Epub 2017 Apr 27.