Liu Zhe, Zheng Yinqiang, Han Xian-Hua
Graduate School of Science and Technology for Innovation, Yamaguchi University, Yamaguchi 753-8511, Japan.
National Institute of Informatics, Tokyo 101-8430, Japan.
Sensors (Basel). 2021 Mar 28;21(7):2348. doi: 10.3390/s21072348.
Hyperspectral image (HSI) super-resolution (SR) is a challenging task due to its ill-posed nature, and has attracted extensive attention by the research community. Previous methods concentrated on leveraging various hand-crafted image priors of a latent high-resolution hyperspectral (HR-HS) image to regularize the degradation model of the observed low-resolution hyperspectral (LR-HS) and HR-RGB images. Different optimization strategies for searching a plausible solution, which usually leads to a limited reconstruction performance, were also exploited. Recently, deep-learning-based methods evolved for automatically learning the abundant image priors in a latent HR-HS image. These methods have made great progress for HS image super resolution. Current deep-learning methods have faced difficulties in designing more complicated and deeper neural network architectures for boosting the performance. They also require large-scale training triplets, such as the LR-HS, HR-RGB, and their corresponding HR-HS images for neural network training. These training triplets significantly limit their applicability to real scenarios. In this work, a deep unsupervised fusion-learning framework for generating a latent HR-HS image using only the observed LR-HS and HR-RGB images without previous preparation of any other training triplets is proposed. Based on the fact that a convolutional neural network architecture is capable of capturing a large number of low-level statistics (priors) of images, the automatic learning of underlying priors of spatial structures and spectral attributes in a latent HR-HS image using only its corresponding degraded observations is promoted. Specifically, the parameter space of a generative neural network used for learning the required HR-HS image to minimize the reconstruction errors of the observations using mathematical relations between data is investigated. Moreover, special convolutional layers for approximating the degradation operations between observations and the latent HR-HS image are specifically to construct an end-to-end unsupervised learning framework for HS image super-resolution. Experiments on two benchmark HS datasets, including the CAVE and Harvard, demonstrate that the proposed method can is capable of producing very promising results, even under a large upscaling factor. Furthermore, it can outperform other unsupervised state-of-the-art methods by a large margin, and manifests its superiority and efficiency.
高光谱图像(HSI)超分辨率(SR)由于其不适定性质而成为一项具有挑战性的任务,并已引起研究界的广泛关注。先前的方法集中于利用潜在高分辨率高光谱(HR-HS)图像的各种手工制作的图像先验来正则化观测到的低分辨率高光谱(LR-HS)和HR-RGB图像的退化模型。还采用了不同的优化策略来寻找合理的解决方案,这通常导致有限的重建性能。最近,基于深度学习的方法发展起来,用于自动学习潜在HR-HS图像中丰富的图像先验。这些方法在HS图像超分辨率方面取得了很大进展。当前的深度学习方法在设计更复杂、更深的神经网络架构以提高性能方面面临困难。它们还需要大规模的训练三元组,例如用于神经网络训练的LR-HS、HR-RGB及其相应的HR-HS图像。这些训练三元组显著限制了它们在实际场景中的适用性。在这项工作中,提出了一种深度无监督融合学习框架,该框架仅使用观测到的LR-HS和HR-RGB图像生成潜在的HR-HS图像,而无需事先准备任何其他训练三元组。基于卷积神经网络架构能够捕获图像的大量低级统计信息(先验)这一事实,促进了仅使用其相应的退化观测值自动学习潜在HR-HS图像中空间结构和光谱属性的潜在先验。具体而言,研究了用于学习所需HR-HS图像以使用数据之间的数学关系最小化观测值重建误差的生成神经网络的参数空间。此外,专门用于近似观测值与潜在HR-HS图像之间退化操作的特殊卷积层,专门构建了一个用于HS图像超分辨率的端到端无监督学习框架。在包括CAVE和Harvard在内的两个基准HS数据集上进行的实验表明,所提出的方法即使在大的放大因子下也能够产生非常有前景的结果。此外,它可以在很大程度上优于其他无监督的最新方法,并显示出其优越性和效率。