Montiel-Marín Santiago, Llamazares Ángel, Antunes Miguel, Revenga Pedro A, Bergasa Luis M
Department of Electronics, Universidad de Alcalá, 28805 Alcalá de Henares, Spain.
Sensors (Basel). 2024 Feb 15;24(4):1244. doi: 10.3390/s24041244.
RADARs and cameras have been present in automotives since the advent of ADAS, as they possess complementary strengths and weaknesses but have been underlooked in the context of learning-based methods. In this work, we propose a method to perform object detection in autonomous driving based on a geometrical and sequential sensor fusion of 3+1D RADAR and semantics extracted from camera data through point cloud painting from the perspective view. To achieve this objective, we adapt PointPainting from the LiDAR and camera domains to the sensors mentioned above. We first apply YOLOv8-seg to obtain instance segmentation masks and project their results to the point cloud. As a refinement stage, we design a set of heuristic rules to minimize the propagation of errors from the segmentation to the detection stage. Our pipeline concludes by applying PointPillars as an object detection network to the painted RADAR point cloud. We validate our approach in the novel View of Delft dataset, which includes 3+1D RADAR data sequences in urban environments. Experimental results show that this fusion is also suitable for RADAR and cameras as we obtain a significant improvement over the RADAR-only baseline, increasing mAP from 41.18 to 52.67 (+27.9%).
自高级驾驶辅助系统(ADAS)出现以来,雷达(RADAR)和摄像头就已应用于汽车领域,因为它们各有优缺点,但在基于学习的方法中却被忽视了。在这项工作中,我们提出了一种基于3 + 1D雷达的几何和序列传感器融合以及通过透视视图的点云绘制从相机数据中提取的语义来在自动驾驶中进行目标检测的方法。为实现这一目标,我们将适用于激光雷达(LiDAR)和摄像头领域的点云绘制(PointPainting)方法应用于上述传感器。我们首先应用YOLOv8-seg来获得实例分割掩码,并将其结果投影到点云上。作为一个细化阶段,我们设计了一组启发式规则,以尽量减少从分割到检测阶段的误差传播。我们的流程最后将PointPillars作为目标检测网络应用于绘制后的雷达点云。我们在新颖的代尔夫特视图(View of Delft)数据集中验证了我们的方法,该数据集包括城市环境中的3 + 1D雷达数据序列。实验结果表明,这种融合也适用于雷达和摄像头,因为我们相对于仅使用雷达的基线有显著改进,平均精度均值(mAP)从41.18提高到52.67(+27.9%)。