School of Computer Science, Shaanxi Normal University, Xi'an 710062, China.
School of Mathematics and Computer Application, Shangluo University, Shangluo 726000, China.
Sensors (Basel). 2022 Jul 31;22(15):5735. doi: 10.3390/s22155735.
Deep learning techniques have brought substantial performance gains to remote sensing image classification. Among them, convolutional neural networks (CNN) can extract rich spatial and spectral features from hyperspectral images in a short-range region, whereas graph convolutional networks (GCN) can model middle- and long-range spatial relations (or structural features) between samples on their graph structure. These different features make it possible to classify remote sensing images finely. In addition, hyperspectral images and light detection and ranging (LiDAR) images can provide spatial-spectral information and elevation information of targets on the Earth's surface, respectively. These multi-source remote sensing data can further improve classification accuracy in complex scenes. This paper proposes a classification method for HS and LiDAR data based on a dual-coupled CNN-GCN structure. The model can be divided into a coupled CNN and a coupled GCN. The former employs a weight-sharing mechanism to structurally fuse and simplify the dual CNN models and extracting the spatial features from HS and LiDAR data. The latter first concatenates the HS and LiDAR data to construct a uniform graph structure. Then, the dual GCN models perform structural fusion by sharing the graph structures and weight matrices of some layers to extract their structural information, respectively. Finally, the final hybrid features are fed into a standard classifier for the pixel-level classification task under a unified feature fusion module. Extensive experiments on two real-world hyperspectral and LiDAR data demonstrate the effectiveness and superiority of the proposed method compared to other state-of-the-art baseline methods, such as two-branch CNN and context CNN. In particular, the overall accuracy (99.11%) on Trento achieves the best classification performance reported so far.
深度学习技术为遥感图像分类带来了显著的性能提升。其中,卷积神经网络(CNN)可以在短距离范围内从高光谱图像中提取丰富的空间和光谱特征,而图卷积网络(GCN)可以在其图结构上对样本之间的中长距离空间关系(或结构特征)进行建模。这些不同的特征使得精细地对遥感图像进行分类成为可能。此外,高光谱图像和激光雷达(LiDAR)图像分别可以提供目标在地球表面的空间光谱信息和高程信息。这些多源遥感数据可以进一步提高复杂场景下的分类精度。本文提出了一种基于双耦合 CNN-GCN 结构的 HS 和 LiDAR 数据分类方法。该模型可分为耦合 CNN 和耦合 GCN。前者采用权值共享机制对双 CNN 模型进行结构融合和简化,从 HS 和 LiDAR 数据中提取空间特征。后者首先将 HS 和 LiDAR 数据进行串联,构建统一的图结构。然后,双 GCN 模型通过共享某些层的图结构和权值矩阵,分别对其结构信息进行结构融合。最后,将最终的混合特征输入到统一的特征融合模块中的标准分类器中,用于像素级分类任务。在两个真实的高光谱和 LiDAR 数据集上的广泛实验表明,与其他最先进的基线方法(如双分支 CNN 和上下文 CNN)相比,所提出的方法具有有效性和优越性。特别是在特伦托数据集上,达到了迄今为止报告的最佳分类性能,总体准确率(99.11%)。