• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于结构化驾驶场景深度估计的雷达-相机融合网络

Radar-Camera Fusion Network for Depth Estimation in Structured Driving Scenes.

作者信息

Li Shuguang, Yan Jiafu, Chen Haoran, Zheng Ke

机构信息

School of Automation Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China.

School of Mechanical and Electrical Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China.

出版信息

Sensors (Basel). 2023 Aug 31;23(17):7560. doi: 10.3390/s23177560.

DOI:10.3390/s23177560
PMID:37688016
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10490688/
Abstract

Depth estimation is an important part of the perception system in autonomous driving. Current studies often reconstruct dense depth maps from RGB images and sparse depth maps obtained from other sensors. However, existing methods often pay insufficient attention to latent semantic information. Considering the highly structured characteristics of driving scenes, we propose a dual-branch network to predict dense depth maps by fusing radar and RGB images. The driving scene is divided into three parts in the proposed architecture, each predicting a depth map, which is finally merged into one by implementing the fusion strategy in order to make full use of the potential semantic information in the driving scene. In addition, a variant L1 loss function is applied in the training phase, directing the network to focus more on those areas of interest when driving. Our proposed method is evaluated on the nuScenes dataset. Experiments demonstrate its effectiveness in comparison with previous state of the art methods.

摘要

深度估计是自动驾驶感知系统的重要组成部分。当前的研究通常从RGB图像和从其他传感器获得的稀疏深度图重建密集深度图。然而,现有方法往往对潜在语义信息关注不足。考虑到驾驶场景的高度结构化特征,我们提出了一种双分支网络,通过融合雷达和RGB图像来预测密集深度图。在所提出的架构中,驾驶场景被分为三个部分,每个部分预测一个深度图,最后通过实施融合策略将其合并为一个深度图,以便充分利用驾驶场景中的潜在语义信息。此外,在训练阶段应用了一种变体L1损失函数,引导网络在驾驶时更多地关注那些感兴趣的区域。我们提出的方法在nuScenes数据集上进行了评估。实验表明,与先前的最先进方法相比,它是有效的。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d6e/10490688/1bcf9b978fe3/sensors-23-07560-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d6e/10490688/9daa84894722/sensors-23-07560-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d6e/10490688/2fb48ed4511d/sensors-23-07560-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d6e/10490688/9d6d8a5ee46e/sensors-23-07560-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d6e/10490688/c84eece1f576/sensors-23-07560-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d6e/10490688/13e5bf805b88/sensors-23-07560-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d6e/10490688/47004162f1d4/sensors-23-07560-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d6e/10490688/9d43268e5dd4/sensors-23-07560-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d6e/10490688/4ede096922e6/sensors-23-07560-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d6e/10490688/3004449228ac/sensors-23-07560-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d6e/10490688/1bcf9b978fe3/sensors-23-07560-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d6e/10490688/9daa84894722/sensors-23-07560-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d6e/10490688/2fb48ed4511d/sensors-23-07560-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d6e/10490688/9d6d8a5ee46e/sensors-23-07560-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d6e/10490688/c84eece1f576/sensors-23-07560-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d6e/10490688/13e5bf805b88/sensors-23-07560-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d6e/10490688/47004162f1d4/sensors-23-07560-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d6e/10490688/9d43268e5dd4/sensors-23-07560-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d6e/10490688/4ede096922e6/sensors-23-07560-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d6e/10490688/3004449228ac/sensors-23-07560-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/5d6e/10490688/1bcf9b978fe3/sensors-23-07560-g010.jpg

相似文献

1
Radar-Camera Fusion Network for Depth Estimation in Structured Driving Scenes.用于结构化驾驶场景深度估计的雷达-相机融合网络
Sensors (Basel). 2023 Aug 31;23(17):7560. doi: 10.3390/s23177560.
2
Learning to Reconstruct and Understand Indoor Scenes From Sparse Views.从稀疏视图中学习重建和理解室内场景。
IEEE Trans Image Process. 2020 Apr 14. doi: 10.1109/TIP.2020.2986712.
3
Guided Depth Completion with Instance Segmentation Fusion in Autonomous Driving Applications.自动驾驶应用中的实例分割融合引导深度补全。
Sensors (Basel). 2022 Dec 7;22(24):9578. doi: 10.3390/s22249578.
4
A Transformer-Based Image-Guided Depth-Completion Model with Dual-Attention Fusion Module.一种基于Transformer的具有双注意力融合模块的图像引导深度补全模型。
Sensors (Basel). 2024 Sep 27;24(19):6270. doi: 10.3390/s24196270.
5
RCRFNet: Enhancing Object Detection with Self-Supervised Radar-Camera Fusion and Open-Set Recognition.RCRFNet:通过自监督雷达-相机融合和开放集识别增强目标检测
Sensors (Basel). 2024 Jul 24;24(15):4803. doi: 10.3390/s24154803.
6
EPMF: Efficient Perception-Aware Multi-Sensor Fusion for 3D Semantic Segmentation.EPMF:用于3D语义分割的高效感知感知多传感器融合
IEEE Trans Pattern Anal Mach Intell. 2024 Dec;46(12):8258-8273. doi: 10.1109/TPAMI.2024.3402232. Epub 2024 Nov 6.
7
Monocular Depth Estimation from a Fisheye Camera Based on Knowledge Distillation.基于知识蒸馏的鱼眼相机单目深度估计
Sensors (Basel). 2023 Dec 16;23(24):9866. doi: 10.3390/s23249866.
8
Expanding Sparse Radar Depth Based on Joint Bilateral Filter for Radar-Guided Monocular Depth Estimation.基于联合双边滤波器扩展稀疏雷达深度用于雷达引导的单目深度估计
Sensors (Basel). 2024 Mar 14;24(6):1864. doi: 10.3390/s24061864.
9
SFA-MDEN: Semantic-Feature-Aided Monocular Depth Estimation Network Using Dual Branches.SFA-MDEN:基于语义特征辅助的双通道单目深度估计网络。
Sensors (Basel). 2021 Aug 13;21(16):5476. doi: 10.3390/s21165476.
10
Real time object detection using LiDAR and camera fusion for autonomous driving.基于激光雷达和相机融合的自动驾驶实时目标检测。
Sci Rep. 2023 May 17;13(1):8056. doi: 10.1038/s41598-023-35170-z.

引用本文的文献

1
Analysis of Advanced Driver-Assistance Systems for Safe and Comfortable Driving of Motor Vehicles.用于机动车安全舒适驾驶的先进驾驶辅助系统分析
Sensors (Basel). 2024 Sep 26;24(19):6223. doi: 10.3390/s24196223.

本文引用的文献

1
Emergency Response Person Localization and Vital Sign Estimation Using a Semi-Autonomous Robot Mounted SFCW Radar.利用搭载半自主机器人的 SFCW 雷达进行应急响应人员定位和生命体征估计。
IEEE Trans Biomed Eng. 2024 Jun;71(6):1756-1769. doi: 10.1109/TBME.2024.3350789. Epub 2024 May 20.
2
mmPose-NLP: A Natural Language Processing Approach to Precise Skeletal Pose Estimation Using mmWave Radars.毫米波姿态自然语言处理:一种使用毫米波雷达进行精确骨骼姿态估计的自然语言处理方法。
IEEE Trans Neural Netw Learn Syst. 2023 Nov;34(11):8418-8429. doi: 10.1109/TNNLS.2022.3151101. Epub 2023 Oct 27.
3
Learning Depth with Convolutional Spatial Propagation Network.
基于卷积空间传播网络的深度学习
IEEE Trans Pattern Anal Mach Intell. 2020 Oct;42(10):2361-2379. doi: 10.1109/TPAMI.2019.2947374. Epub 2019 Oct 15.
4
Joint Task-Recursive Learning for RGB-D Scene Understanding.用于RGB-D场景理解的联合任务递归学习
IEEE Trans Pattern Anal Mach Intell. 2020 Oct;42(10):2608-2623. doi: 10.1109/TPAMI.2019.2926728. Epub 2019 Jul 10.
5
Deep Ordinal Regression Network for Monocular Depth Estimation.用于单目深度估计的深度序数回归网络
Proc IEEE Comput Soc Conf Comput Vis Pattern Recognit. 2018 Jun;2018:2002-2011. doi: 10.1109/CVPR.2018.00214. Epub 2018 Dec 17.