Zhang Kun, Chen Rui, Peng Zidong, Zhu Yawei, Wang Xiaohong
College of Information Science and Engineering, Hebei University of Science and Technology, Shijiazhuang 050018, China.
College of International Education, Guangxi University of Science and Technology, Liuzhou 545006, China.
Sensors (Basel). 2023 Oct 9;23(19):8338. doi: 10.3390/s23198338.
In interpreting a scene for numerous applications, including autonomous driving and robotic navigation, semantic segmentation is crucial. Compared to single-modal data, multi-modal data allow us to extract a richer set of features, which is the benefit of improving segmentation accuracy and effect. We propose a point cloud semantic segmentation method, and a fusion graph convolutional network (FGCN) which extracts the semantic information of each point involved in the two-modal data of images and point clouds. The two-channel k-nearest neighbors (KNN) module of the FGCN was created to address the issue of the feature extraction's poor efficiency by utilizing picture data. Notably, the FGCN utilizes the spatial attention mechanism to better distinguish more important features and fuses multi-scale features to enhance the generalization capability of the network and increase the accuracy of the semantic segmentation. In the experiment, a self-made semantic segmentation KITTI (SSKIT) dataset was made for the fusion effect. The mean intersection over union (MIoU) of the SSKIT can reach 88.06%. As well as the public datasets, the S3DIS showed that our method can enhance data features and outperform other methods: the MIoU of the S3DIS can reach up to 78.55%. The segmentation accuracy is significantly improved compared with the existing methods, which verifies the effectiveness of the improved algorithms.
在为包括自动驾驶和机器人导航在内的众多应用场景解读场景时,语义分割至关重要。与单模态数据相比,多模态数据使我们能够提取更丰富的特征集,这有利于提高分割精度和效果。我们提出了一种点云语义分割方法以及一种融合图卷积网络(FGCN),该网络可提取图像和点云这两种模态数据中每个点的语义信息。FGCN的双通道k近邻(KNN)模块旨在通过利用图像数据来解决特征提取效率低下的问题。值得注意的是,FGCN利用空间注意力机制更好地区分更重要的特征,并融合多尺度特征以增强网络的泛化能力并提高语义分割的准确性。在实验中,为融合效果制作了一个自制的语义分割KITTI(SSKIT)数据集。SSKIT的平均交并比(MIoU)可达88.06%。在公共数据集方面,S3DIS表明我们的方法可以增强数据特征并优于其他方法:S3DIS的MIoU可达78.55%。与现有方法相比,分割精度有显著提高,这验证了改进算法的有效性。