Yan Wanzi, Zhang Yidong, Xue Minti, Zhu Zhencai, Lu Hao, Zhang Xin, Tang Wei, Xing Keke
School of Mines, China University of Mining and Technology, Xuzhou 221116, China.
State Key Laboratory for Fine Exploration and Intelligent Development of Coal Resources, School of Mines, China University of Mining and Technology, Xuzhou 221116, China.
Sensors (Basel). 2025 Aug 20;25(16):5185. doi: 10.3390/s25165185.
Environmental perception is crucial for achieving autonomous driving of auxiliary haulage vehicles in underground coal mines. The complex underground environment and working conditions, such as dust pollution, uneven lighting, and sensor data abnormalities, pose challenges to multimodal fusion perception. These challenges include: (1) the lack of a reasonable and effective method for evaluating the reliability of different modality data; (2) the absence of in-depth fusion methods for different modality data that can handle sensor failures; and (3) the lack of a multimodal dataset for underground coal mines to support model training. To address these issues, this paper proposes a coal mine underground BEV multiscale-enhanced fusion perception model based on dynamic weight adjustment. First, camera and LiDAR modality data are uniformly mapped into BEV space to achieve multimodal feature alignment. Then, a Mixture of Experts-Fuzzy Logic Inference Module (MoE-FLIM) is designed to infer weights for different modality data based on BEV feature dimensions. Next, a Pyramid Multiscale Feature Enhancement and Fusion Module (PMS-FFEM) is introduced to ensure the model's perception performance in the event of sensor data abnormalities. Lastly, a multimodal dataset for underground coal mines is constructed to provide support for model training and testing in real-world scenarios. Experimental results show that the proposed method demonstrates good accuracy and stability in object-detection tasks in coal mine underground environments, maintaining high detection performance, especially in typical complex scenes such as low light and dust fog.
环境感知对于实现煤矿井下辅助运输车辆的自动驾驶至关重要。复杂的井下环境和工作条件,如粉尘污染、光照不均以及传感器数据异常等,给多模态融合感知带来了挑战。这些挑战包括:(1)缺乏合理有效的方法来评估不同模态数据的可靠性;(2)缺少能够处理传感器故障的针对不同模态数据的深度融合方法;(3)缺乏用于煤矿井下支持模型训练的多模态数据集。为了解决这些问题,本文提出了一种基于动态权重调整的煤矿井下鸟瞰图多尺度增强融合感知模型。首先,将相机和激光雷达模态数据统一映射到鸟瞰图空间,以实现多模态特征对齐。然后,设计了一个专家混合-模糊逻辑推理模块(MoE-FLIM),基于鸟瞰图特征维度推断不同模态数据的权重。接下来,引入了一个金字塔多尺度特征增强与融合模块(PMS-FFEM),以确保在传感器数据异常情况下模型的感知性能。最后,构建了一个煤矿井下多模态数据集,为实际场景中的模型训练和测试提供支持。实验结果表明,所提方法在煤矿井下环境的目标检测任务中表现出良好的准确性和稳定性,保持了较高的检测性能,尤其是在低光照和粉尘雾等典型复杂场景中。