• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

Infra-3DRC融合网络:用于交通用户检测的路边安装RGB单目摄像头与三维汽车雷达的深度融合

Infra-3DRC-FusionNet: Deep Fusion of Roadside Mounted RGB Mono Camera and Three-Dimensional Automotive Radar for Traffic User Detection.

作者信息

Agrawal Shiva, Bhanderi Savankumar, Elger Gordon

机构信息

Institute of Innovative Mobility (IIMo), Technische Hochschule Ingolstadt, 85049 Ingolstadt, Germany.

Fraunhofer IVI, Applied Center Connected Mobility and Infrastructure, 01069 Dresden, Germany.

出版信息

Sensors (Basel). 2025 May 29;25(11):3422. doi: 10.3390/s25113422.

DOI:10.3390/s25113422
PMID:40968993
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12158296/
Abstract

Mono RGB cameras and automotive radar sensors provide a complementary information set that makes them excellent candidates for sensor data fusion to obtain robust traffic user detection. This has been widely used in the vehicle domain and recently introduced in roadside-mounted smart infrastructure-based road user detection. However, the performance of the most commonly used late fusion methods often degrades when the camera fails to detect road users in adverse environmental conditions. The solution is to fuse the data using deep neural networks at the early stage of the fusion pipeline to use the complete data provided by both sensors. Research has been carried out in this area, but is limited to vehicle-based sensor setups. Hence, this work proposes a novel deep neural network to jointly fuse RGB mono-camera images and 3D automotive radar point cloud data to obtain enhanced traffic user detection for the roadside-mounted smart infrastructure setup. Projected radar points are first used to generate anchors in image regions with a high likelihood of road users, including areas not visible to the camera. These anchors guide the prediction of 2D bounding boxes, object categories, and confidence scores. Valid detections are then used to segment radar points by instance, and the results are post-processed to produce final road user detections in the ground plane. The trained model is evaluated for different light and weather conditions using ground truth data from a lidar sensor. It provides a precision of 92%, recall of 78%, and F1-score of 85%. The proposed deep fusion methodology has 33%, 6%, and 21% absolute improvement in precision, recall, and F1-score, respectively, compared to object-level spatial fusion output.

摘要

单目RGB相机和汽车雷达传感器提供了一组互补的信息,这使得它们成为传感器数据融合以获得强大的交通用户检测能力的理想选择。这已在车辆领域广泛应用,最近也被引入基于路边安装的智能基础设施的道路用户检测中。然而,当相机在恶劣环境条件下无法检测到道路用户时,最常用的后期融合方法的性能往往会下降。解决方案是在融合管道的早期阶段使用深度神经网络融合数据,以利用两个传感器提供的完整数据。这一领域已经开展了相关研究,但仅限于基于车辆的传感器设置。因此,这项工作提出了一种新颖的深度神经网络,用于联合融合RGB单目相机图像和3D汽车雷达点云数据,以增强基于路边安装的智能基础设施设置的交通用户检测能力。首先使用投影雷达点在道路用户可能性高的图像区域生成锚点,包括相机不可见的区域。这些锚点指导二维边界框、物体类别和置信度得分的预测。然后使用有效的检测结果按实例分割雷达点,并对结果进行后处理,以在地面平面上生成最终的道路用户检测结果。使用来自激光雷达传感器的地面真值数据对训练好的模型在不同光照和天气条件下进行评估。它的精度为92%,召回率为78%,F1分数为85%。与目标级空间融合输出相比,所提出的深度融合方法在精度、召回率和F1分数上分别有33%、6%和21%的绝对提升。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/522f/12158296/96b51c8ddbf1/sensors-25-03422-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/522f/12158296/fb9b2fbf2acd/sensors-25-03422-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/522f/12158296/be9e4564fba4/sensors-25-03422-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/522f/12158296/5ef90f8e7cd3/sensors-25-03422-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/522f/12158296/2f8680c6253d/sensors-25-03422-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/522f/12158296/e8ba3d4a56bb/sensors-25-03422-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/522f/12158296/5194ff20d0c0/sensors-25-03422-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/522f/12158296/497562c515f5/sensors-25-03422-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/522f/12158296/493b43ec82be/sensors-25-03422-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/522f/12158296/96b51c8ddbf1/sensors-25-03422-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/522f/12158296/fb9b2fbf2acd/sensors-25-03422-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/522f/12158296/be9e4564fba4/sensors-25-03422-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/522f/12158296/5ef90f8e7cd3/sensors-25-03422-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/522f/12158296/2f8680c6253d/sensors-25-03422-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/522f/12158296/e8ba3d4a56bb/sensors-25-03422-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/522f/12158296/5194ff20d0c0/sensors-25-03422-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/522f/12158296/497562c515f5/sensors-25-03422-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/522f/12158296/493b43ec82be/sensors-25-03422-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/522f/12158296/96b51c8ddbf1/sensors-25-03422-g009.jpg

相似文献

1
Infra-3DRC-FusionNet: Deep Fusion of Roadside Mounted RGB Mono Camera and Three-Dimensional Automotive Radar for Traffic User Detection.Infra-3DRC融合网络:用于交通用户检测的路边安装RGB单目摄像头与三维汽车雷达的深度融合
Sensors (Basel). 2025 May 29;25(11):3422. doi: 10.3390/s25113422.
2
Prescription of Controlled Substances: Benefits and Risks管制药品的处方:益处与风险
3
Short-Term Memory Impairment短期记忆障碍
4
EPCNet: Implementing an 'Artificial Fovea' for More Efficient Monitoring Using the Sensor Fusion of an Event-Based and a Frame-Based Camera.EPCNet:利用基于事件的相机和基于帧的相机的传感器融合实现“人工中央凹”以进行更高效的监测。
Sensors (Basel). 2025 Jul 22;25(15):4540. doi: 10.3390/s25154540.
5
T360Fusion: Temporal 360 Multimodal Fusion for 3D Object Detection via Transformers.T360融合:通过Transformer实现用于3D目标检测的时域360多模态融合
Sensors (Basel). 2025 Aug 8;25(16):4902. doi: 10.3390/s25164902.
6
Aspects of Genetic Diversity, Host Specificity and Public Health Significance of Single-Celled Intestinal Parasites Commonly Observed in Humans and Mostly Referred to as 'Non-Pathogenic'.人类常见且大多被称为“非致病性”的单细胞肠道寄生虫的遗传多样性、宿主特异性及公共卫生意义
APMIS. 2025 Sep;133(9):e70036. doi: 10.1111/apm.70036.
7
Anterior Approach Total Ankle Arthroplasty with Patient-Specific Cut Guides.使用患者特异性截骨导向器的前路全踝关节置换术。
JBJS Essent Surg Tech. 2025 Aug 15;15(3). doi: 10.2106/JBJS.ST.23.00027. eCollection 2025 Jul-Sep.
8
Integrated neural network framework for multi-object detection and recognition using UAV imagery.用于使用无人机图像进行多目标检测与识别的集成神经网络框架。
Front Neurorobot. 2025 Jul 30;19:1643011. doi: 10.3389/fnbot.2025.1643011. eCollection 2025.
9
Comparison of Two Modern Survival Prediction Tools, SORG-MLA and METSSS, in Patients With Symptomatic Long-bone Metastases Who Underwent Local Treatment With Surgery Followed by Radiotherapy and With Radiotherapy Alone.两种现代生存预测工具 SORG-MLA 和 METSSS 在接受手术联合放疗和单纯放疗治疗有症状长骨转移患者中的比较。
Clin Orthop Relat Res. 2024 Dec 1;482(12):2193-2208. doi: 10.1097/CORR.0000000000003185. Epub 2024 Jul 23.
10
Regional cerebral blood flow single photon emission computed tomography for detection of Frontotemporal dementia in people with suspected dementia.用于检测疑似痴呆患者额颞叶痴呆的局部脑血流单光子发射计算机断层扫描
Cochrane Database Syst Rev. 2015 Jun 23;2015(6):CD010896. doi: 10.1002/14651858.CD010896.pub2.

本文引用的文献

1
MmWave Radar and Vision Fusion for Object Detection in Autonomous Driving: A Review.毫米波雷达与视觉融合在自动驾驶中的目标检测:综述。
Sensors (Basel). 2022 Mar 25;22(7):2542. doi: 10.3390/s22072542.
2
Spatial Attention Fusion for Obstacle Detection Using MmWave Radar and Vision Sensor.基于毫米波雷达和视觉传感器的障碍物检测的空间注意力融合。
Sensors (Basel). 2020 Feb 11;20(4):956. doi: 10.3390/s20040956.
3
Sensor Technologies for Intelligent Transportation Systems.智能交通系统的传感器技术
Sensors (Basel). 2018 Apr 16;18(4):1212. doi: 10.3390/s18041212.
4
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks.更快的 R-CNN:基于区域建议网络的实时目标检测。
IEEE Trans Pattern Anal Mach Intell. 2017 Jun;39(6):1137-1149. doi: 10.1109/TPAMI.2016.2577031. Epub 2016 Jun 6.