Wang Xinying, Wu Yingdan, Ming Yang, Lv Hui
School of Science, Hubei University of Technology, No. 28 Nanli Road, Wuhan 430068, China.
Hubei Collaborative Innovation Centre for High-Efficient Utilization of Solar Energy, Hubei University of Technology, No. 28 Nanli Road, Wuhan 430068, China.
Sensors (Basel). 2020 Feb 19;20(4):1142. doi: 10.3390/s20041142.
Due to increasingly complex factors of image degradation, inferring high-frequency details of remote sensing imagery is more difficult compared to ordinary digital photos. This paper proposes an adaptive multi-scale feature fusion network (AMFFN) for remote sensing image super-resolution. Firstly, the features are extracted from the original low-resolution image. Then several adaptive multi-scale feature extraction (AMFE) modules, the squeeze-and-excited and adaptive gating mechanisms are adopted for feature extraction and fusion. Finally, the sub-pixel convolution method is used to reconstruct the high-resolution image. Experiments are performed on three datasets, the key characteristics, such as the number of AMFEs and the gating connection way are studied, and super-resolution of remote sensing imagery of different scale factors are qualitatively and quantitatively analyzed. The results show that our method outperforms the classic methods, such as Super-Resolution Convolutional Neural Network(SRCNN), Efficient Sub-Pixel Convolutional Network (ESPCN), and multi-scale residual CNN(MSRN).
由于图像退化因素日益复杂,与普通数码照片相比,推断遥感图像的高频细节更加困难。本文提出一种用于遥感图像超分辨率的自适应多尺度特征融合网络(AMFFN)。首先,从原始低分辨率图像中提取特征。然后采用若干自适应多尺度特征提取(AMFE)模块、挤压与激励以及自适应门控机制进行特征提取和融合。最后,使用子像素卷积方法重建高分辨率图像。在三个数据集上进行了实验,研究了诸如AMFE数量和门控连接方式等关键特征,并对不同比例因子的遥感图像超分辨率进行了定性和定量分析。结果表明,我们的方法优于经典方法,如超分辨率卷积神经网络(SRCNN)、高效子像素卷积网络(ESPCN)和多尺度残差卷积神经网络(MSRN)。