Jang Harin, Kim Taehyun, Ahn Kyungjae, Jeon Soo, Kang Yeonsik
Graduate School of Automotive Engineering, Kookmin University, 77 Jeongneung-ro, Seongbuk-gu, Seoul 02707, Republic of Korea.
Department of Mechanical and Mechatronics Engineering, University of Waterloo, 200 University Avenue West, Waterloo, ON N2L 3G1, Canada.
Sensors (Basel). 2024 Apr 29;24(9):2828. doi: 10.3390/s24092828.
In the field of robotics and autonomous driving, dynamic occupancy grid maps (DOGMs) are typically used to represent the position and velocity information of objects. Although three-dimensional light detection and ranging (LiDAR) sensor-based DOGMs have been actively researched, they have limitations, as they cannot classify types of objects. Therefore, in this study, a deep learning-based camera-LiDAR sensor fusion technique is employed as input to DOGMs. Consequently, not only the position and velocity information of objects but also their class information can be updated, expanding the application areas of DOGMs. Moreover, unclassified LiDAR point measurements contribute to the formation of a map of the surrounding environment, improving the reliability of perception by registering objects that were not classified by deep learning. To achieve this, we developed update rules on the basis of the Dempster-Shafer evidence theory, incorporating class information and the uncertainty of objects occupying grid cells. Furthermore, we analyzed the accuracy of the velocity estimation using two update models. One assigns the occupancy probability only to the edges of the oriented bounding box, whereas the other assigns the occupancy probability to the entire area of the box. The performance of the developed perception technique is evaluated using the public nuScenes dataset. The developed DOGM with object class information will help autonomous vehicles to navigate in complex urban driving environments by providing them with rich information, such as the class and velocity of nearby obstacles.
在机器人技术和自动驾驶领域,动态占用网格地图(DOGM)通常用于表示物体的位置和速度信息。尽管基于三维激光雷达(LiDAR)传感器的DOGM已得到积极研究,但它们存在局限性,因为无法对物体类型进行分类。因此,在本研究中,一种基于深度学习的摄像头 - LiDAR传感器融合技术被用作DOGM的输入。结果,不仅物体的位置和速度信息,而且其类别信息都可以更新,从而扩展了DOGM的应用领域。此外,未分类的LiDAR点测量有助于形成周围环境地图,通过对深度学习未分类的物体进行配准来提高感知的可靠性。为实现这一点,我们基于Dempster - Shafer证据理论开发了更新规则,纳入了类别信息和占据网格单元的物体的不确定性。此外,我们使用两种更新模型分析了速度估计的准确性。一种仅将占用概率分配给定向包围盒的边缘,而另一种将占用概率分配给包围盒的整个区域。使用公开的nuScenes数据集评估所开发感知技术的性能。所开发的带有物体类别信息的DOGM将通过为自动驾驶车辆提供丰富信息(如附近障碍物的类别和速度)来帮助它们在复杂的城市驾驶环境中导航。