Wang Yan, Yang Ling, Liu Xinzhan, Yan Pengfei
College of Geography and Environmental Science, Henan University, Kaifeng, China.
Key Laboratory of Geospatial Technology for Middle and Lower Yellow River Regions, Ministry of Education, Kaifeng, China.
Sci Rep. 2024 Apr 27;14(1):9716. doi: 10.1038/s41598-024-60375-1.
High-precision and high-efficiency Semantic segmentation of high-resolution remote sensing images is a challenge. Existing models typically require a significant amount of training data to achieve good classification results and have numerous training parameters. A novel model called MST-DeepLabv3+ was suggested in this paper for remote sensing image classification. It's based on the DeepLabv3+ and can produce better results with fewer train parameters. MST-DeepLabv3+ made three improvements: (1) Reducing the number of model parameters by substituting MobileNetV2 for the Xception in the DeepLabv3+'s backbone network. (2) Adding the attention mechanism module SENet to increase the precision of semantic segmentation. (3) Increasing Transfer Learning to enhance the model's capacity to recognize features, and raise the segmentation accuracy. MST-DeepLabv3+ was tested on international society for photogrammetry and remote sensing (ISPRS) dataset, Gaofen image dataset (GID), and practically applied to the Taikang cultivated land dataset. On the ISPRS dataset, the mean intersection over union (MIoU), overall accuracy (OA), Precision, Recall, and F1-score are 82.47%, 92.13%, 90.34%, 90.12%, and 90.23%, respectively. On the GID dataset, these values are 73.44%, 85.58%, 84.10%, 84.86%, and 84.48%, respectively. The results were as high as 90.77%, 95.47%, 95.28%, 95.02%, and 95.15% on the Taikang cultivated land dataset. The experimental results indicate that MST-DeepLabv3+ effectively improves the accuracy of semantic segmentation of remote sensing images, recognizes the edge information with more completeness, and significantly reduces the parameter size.
高分辨率遥感图像的高精度、高效率语义分割是一项挑战。现有模型通常需要大量训练数据才能取得良好的分类结果,并且具有众多训练参数。本文提出了一种名为MST-DeepLabv3+的新型模型用于遥感图像分类。它基于DeepLabv3+,可以用更少的训练参数产生更好的结果。MST-DeepLabv3+进行了三项改进:(1)在DeepLabv3+的骨干网络中用MobileNetV2替换Xception,以减少模型参数数量。(2)添加注意力机制模块SENet以提高语义分割的精度。(3)增加迁移学习以增强模型识别特征的能力,并提高分割精度。MST-DeepLabv3+在国际摄影测量与遥感学会(ISPRS)数据集、高分图像数据集(GID)上进行了测试,并实际应用于太康耕地数据集。在ISPRS数据集上,平均交并比(MIoU)、总体准确率(OA)、精确率、召回率和F1分数分别为82.47%、92.13%、90.34%、90.12%和90.23%。在GID数据集上,这些值分别为73.44%、85.58%、84.10%、84.86%和84.48%。在太康耕地数据集上,结果分别高达90.77%、95.47%、95.28%、95.02%和95.15%。实验结果表明,MST-DeepLabv3+有效地提高了遥感图像语义分割的精度,更完整地识别边缘信息,并显著减小了参数规模。