• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于深度学习的具有深度信息细化的单目 3D 目标检测。

Deep Learning-Based Monocular 3D Object Detection with Refinement of Depth Information.

机构信息

Changchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun 130033, China.

University of Chinese Academy of Sciences, Beijing 100049, China.

出版信息

Sensors (Basel). 2022 Mar 28;22(7):2576. doi: 10.3390/s22072576.

DOI:10.3390/s22072576
PMID:35408191
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9003335/
Abstract

Recently, the research on monocular 3D target detection based on pseudo-LiDAR data has made some progress. In contrast to LiDAR-based algorithms, the robustness of pseudo-LiDAR methods is still inferior. After conducting in-depth experiments, we realized that the main limitations are due to the inaccuracy of the target position and the uncertainty in the depth distribution of the foreground target. These two problems arise from the inaccurate depth estimation. To deal with the aforementioned problems, we propose two innovative solutions. The first is a novel method based on joint image segmentation and geometric constraints, used to predict the target depth and provide the depth prediction confidence measure. The predicted target depth is fused with the overall depth of the scene and results in the optimal target position. For the second, we utilize the target scale, normalized with the Gaussian function, as a priori information. The uncertainty of depth distribution, which can be visualized as long-tail noise, is reduced. With the refined depth information, we convert the optimized depth map into the point cloud representation, called a pseudo-LiDAR point cloud. Finally, we input the pseudo-LiDAR point cloud to the LiDAR-based algorithm to detect the 3D target. We conducted extensive experiments on the challenging KITTI dataset. The results demonstrate that our proposed framework outperforms various state-of-the-art methods by more than 12.37% and 5.34% on the easy and hard settings of the KITTI validation subset, respectively. On the KITTI test set, our framework also outperformed state-of-the-art methods by 5.1% and 1.76% on the easy and hard settings, respectively.

摘要

最近,基于伪激光雷达数据的单目 3D 目标检测研究取得了一些进展。与基于激光雷达的算法相比,伪激光雷达方法的鲁棒性仍然较差。经过深入实验,我们意识到主要的限制是由于目标位置的不准确性和前景目标深度分布的不确定性。这两个问题源于深度估计的不准确性。为了解决上述问题,我们提出了两种创新的解决方案。第一个是一种基于联合图像分割和几何约束的新方法,用于预测目标深度并提供深度预测置信度度量。预测的目标深度与场景的整体深度融合,得出最优的目标位置。对于第二个问题,我们利用目标尺度,归一化为高斯函数,作为先验信息。深度分布的不确定性,可视为长尾噪声,得到降低。利用细化后的深度信息,我们将优化后的深度图转换为点云表示,称为伪激光雷达点云。最后,我们将伪激光雷达点云输入基于激光雷达的算法中,以检测 3D 目标。我们在具有挑战性的 KITTI 数据集上进行了广泛的实验。结果表明,我们提出的框架在 KITTI 验证子集的简单和困难设置上分别比各种最先进的方法高出 12.37%和 5.34%。在 KITTI 测试集上,我们的框架在简单和困难设置上也分别比最先进的方法高出 5.1%和 1.76%。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2e2d/9003335/5d3b54586083/sensors-22-02576-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2e2d/9003335/1d129ac52b0d/sensors-22-02576-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2e2d/9003335/1ffdf1641653/sensors-22-02576-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2e2d/9003335/a6888cb0549f/sensors-22-02576-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2e2d/9003335/710f988fb322/sensors-22-02576-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2e2d/9003335/0b2b52794e2d/sensors-22-02576-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2e2d/9003335/f50b8ea1a230/sensors-22-02576-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2e2d/9003335/2c7cd3023fdf/sensors-22-02576-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2e2d/9003335/657867bdad62/sensors-22-02576-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2e2d/9003335/5d3b54586083/sensors-22-02576-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2e2d/9003335/1d129ac52b0d/sensors-22-02576-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2e2d/9003335/1ffdf1641653/sensors-22-02576-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2e2d/9003335/a6888cb0549f/sensors-22-02576-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2e2d/9003335/710f988fb322/sensors-22-02576-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2e2d/9003335/0b2b52794e2d/sensors-22-02576-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2e2d/9003335/f50b8ea1a230/sensors-22-02576-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2e2d/9003335/2c7cd3023fdf/sensors-22-02576-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2e2d/9003335/657867bdad62/sensors-22-02576-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2e2d/9003335/5d3b54586083/sensors-22-02576-g009.jpg

相似文献

1
Deep Learning-Based Monocular 3D Object Detection with Refinement of Depth Information.基于深度学习的具有深度信息细化的单目 3D 目标检测。
Sensors (Basel). 2022 Mar 28;22(7):2576. doi: 10.3390/s22072576.
2
MonoDCN: Monocular 3D object detection based on dynamic convolution.MonoDCN:基于动态卷积的单目三维目标检测。
PLoS One. 2022 Oct 4;17(10):e0275438. doi: 10.1371/journal.pone.0275438. eCollection 2022.
3
Transfer Learning Based Semantic Segmentation for 3D Object Detection from Point Cloud.基于迁移学习的点云三维目标检测语义分割。
Sensors (Basel). 2021 Jun 8;21(12):3964. doi: 10.3390/s21123964.
4
Point-Guided Contrastive Learning for Monocular 3-D Object Detection.用于单目三维目标检测的点引导对比学习
IEEE Trans Cybern. 2023 Feb;53(2):954-966. doi: 10.1109/TCYB.2021.3090370. Epub 2023 Jan 13.
5
PLIN: A Network for Pseudo-LiDAR Point Cloud Interpolation.PLIN:用于伪激光雷达点云插值的网络。
Sensors (Basel). 2020 Mar 12;20(6):1573. doi: 10.3390/s20061573.
6
Efficient Stereo Depth Estimation for Pseudo-LiDAR: A Self-Supervised Approach Based on Multi-Input ResNet Encoder.基于多输入 ResNet 编码器的伪激光雷达高效立体深度估计:一种自监督方法。
Sensors (Basel). 2023 Feb 2;23(3):1650. doi: 10.3390/s23031650.
7
PTA-Det: Point Transformer Associating Point Cloud and Image for 3D Object Detection.PTA-Det:用于 3D 目标检测的点变换关联点云和图像。
Sensors (Basel). 2023 Mar 17;23(6):3229. doi: 10.3390/s23063229.
8
EPGNet: Enhanced Point Cloud Generation for 3D Object Detection.EPGNet:用于 3D 目标检测的增强点云生成。
Sensors (Basel). 2020 Dec 4;20(23):6927. doi: 10.3390/s20236927.
9
A Survey on Deep-Learning-Based LiDAR 3D Object Detection for Autonomous Driving.基于深度学习的自动驾驶激光雷达 3D 目标检测研究综述。
Sensors (Basel). 2022 Dec 7;22(24):9577. doi: 10.3390/s22249577.
10
Up-Sampling Method for Low-Resolution LiDAR Point Cloud to Enhance 3D Object Detection in an Autonomous Driving Environment.用于增强自动驾驶环境中 3D 目标检测的低分辨率 LiDAR 点云的上采样方法。
Sensors (Basel). 2022 Dec 28;23(1):322. doi: 10.3390/s23010322.

引用本文的文献

1
SDC-Net++: End-to-End Crash Detection and Action Control for Self-Driving Car Deep-IoT-Based System.SDC-Net++:用于基于深度物联网的自动驾驶汽车系统的端到端碰撞检测与动作控制
Sensors (Basel). 2024 Jun 12;24(12):3805. doi: 10.3390/s24123805.
2
Uncertainty Prediction for Monocular 3D Object Detection.单目三维目标检测的不确定性预测。
Sensors (Basel). 2023 Jun 7;23(12):5395. doi: 10.3390/s23125395.

本文引用的文献

1
MDS-Net: Multi-Scale Depth Stratification 3D Object Detection from Monocular Images.MDS-Net:基于单目图像的多尺度深度分层 3D 目标检测
Sensors (Basel). 2022 Aug 18;22(16):6197. doi: 10.3390/s22166197.
2
Multi-Scale Spatial Attention-Guided Monocular Depth Estimation With Semantic Enhancement.具有语义增强的多尺度空间注意力引导单目深度估计
IEEE Trans Image Process. 2021;30:8811-8822. doi: 10.1109/TIP.2021.3120670. Epub 2021 Oct 27.
3
Monocular Depth Estimation with Joint Attention Feature Distillation and Wavelet-Based Loss Function.
基于联合注意特征提取和基于小波的损失函数的单目深度估计。
Sensors (Basel). 2020 Dec 24;21(1):54. doi: 10.3390/s21010054.
4
Deep Learning for LiDAR Point Clouds in Autonomous Driving: A Review.自动驾驶中用于激光雷达点云的深度学习:综述
IEEE Trans Neural Netw Learn Syst. 2021 Aug;32(8):3412-3432. doi: 10.1109/TNNLS.2020.3015992. Epub 2021 Aug 3.
5
Deep Learning-Based Monocular Depth Estimation Methods-A State-of-the-Art Review.基于深度学习的单目深度估计方法——最新综述。
Sensors (Basel). 2020 Apr 16;20(8):2272. doi: 10.3390/s20082272.
6
Deep Ordinal Regression Network for Monocular Depth Estimation.用于单目深度估计的深度序数回归网络
Proc IEEE Comput Soc Conf Comput Vis Pattern Recognit. 2018 Jun;2018:2002-2011. doi: 10.1109/CVPR.2018.00214. Epub 2018 Dec 17.