Shang Ruibo, Hoffer-Hawlik Kevin, Wang Fei, Situ Guohai, Luke Geoffrey P
Opt Express. 2021 May 10;29(10):15239-15254. doi: 10.1364/OE.424165.
Deep learning (DL) is a powerful tool in computational imaging for many applications. A common strategy is to use a preprocessor to reconstruct a preliminary image as the input to a neural network to achieve an optimized image. Usually, the preprocessor incorporates knowledge of the physics priors in the imaging model. One outstanding challenge, however, is errors that arise from imperfections in the assumed model. Model mismatches degrade the quality of the preliminary image and therefore affect the DL predictions. Another main challenge is that many imaging inverse problems are ill-posed and the networks are over-parameterized; DL networks have flexibility to extract features from the data that are not directly related to the imaging model. This can lead to suboptimal training and poorer image reconstruction results. To solve these challenges, a two-step training DL (TST-DL) framework is proposed for computational imaging without physics priors. First, a single fully-connected layer (FCL) is trained to directly learn the inverse model with the raw measurement data as the inputs and the images as the outputs. Then, this pre-trained FCL is fixed and concatenated with an un-trained deep convolutional network with a U-Net architecture for a second-step training to optimize the output image. This approach has the advantage that does not rely on an accurate representation of the imaging physics since the first-step training directly learns the inverse model. Furthermore, the TST-DL approach mitigates network over-parameterization by separately training the FCL and U-Net. We demonstrate this framework using a linear single-pixel camera imaging model. The results are quantitatively compared with those from other frameworks. The TST-DL approach is shown to perform comparable to approaches which incorporate perfect knowledge of the imaging model, to be robust to noise and model ill-posedness, and to be more robust to model mismatch than approaches which incorporate imperfect knowledge of the imaging model. Furthermore, TST-DL yields better results than end-to-end training while suffering from less overfitting. Overall, this TST-DL framework is a flexible approach for image reconstruction without physics priors, applicable to diverse computational imaging systems.
深度学习(DL)在计算成像的诸多应用中是一种强大的工具。一种常见策略是使用预处理器来重建初步图像,作为神经网络的输入以获得优化图像。通常,预处理器会纳入成像模型中的物理先验知识。然而,一个突出的挑战是假设模型中的不完美所产生的误差。模型不匹配会降低初步图像的质量,进而影响深度学习的预测。另一个主要挑战是,许多成像逆问题是不适定的,并且网络参数过多;深度学习网络有从数据中提取与成像模型不直接相关特征的灵活性。这可能导致次优训练和较差的图像重建结果。为了解决这些挑战,提出了一种用于无物理先验知识的计算成像的两步训练深度学习(TST-DL)框架。首先,训练一个单全连接层(FCL),以原始测量数据为输入、图像为输出直接学习逆模型。然后,固定这个预训练的FCL,并将其与具有U-Net架构的未训练深度卷积网络连接起来进行第二步训练,以优化输出图像。这种方法的优点是不依赖成像物理的精确表示,因为第一步训练直接学习逆模型。此外,TST-DL方法通过分别训练FCL和U-Net减轻了网络参数过多的问题。我们使用线性单像素相机成像模型展示了这个框架。将结果与其他框架的结果进行了定量比较。结果表明,TST-DL方法的性能与纳入成像模型完美知识的方法相当,对噪声和模型不适定性具有鲁棒性,并且比纳入成像模型不完美知识的方法对模型不匹配更具鲁棒性。此外,TST-DL比端到端训练产生更好的结果,同时过拟合程度更低。总体而言,这个TST-DL框架是一种用于无物理先验知识的图像重建的灵活方法,适用于各种计算成像系统。