Li Jinxing, Guo Xiaobao, Lu Guangming, Zhang Bob, Xu Yong, Wu Feng, Zhang David
IEEE Trans Image Process. 2020 Mar 2. doi: 10.1109/TIP.2020.2976190.
In this paper, a novel deep network is proposed for multi-focus image fusion, named Deep Regression Pair Learning (DRPL). In contrast to existing deep fusion methods which divide the input image into small patches and apply a classifier to judge whether the patch is in focus or not, DRPL directly converts the whole image into a binary mask without any patch operation, subsequently tackling the difficulty of the blur level estimation around the focused/defocused boundary. Simultaneously, a pair learning strategy, which takes a pair of complementary source images as inputs and generates two corresponding binary masks, is introduced into the model, greatly imposing the complementary constraint on each pair and making a large contribution to the performance improvement. Furthermore, as the edge or gradient does exist in the focus part while there is no similar property for the defocus part, we also embed a gradient loss to ensure the generated image to be all-in-focus. Then the structural similarity index (SSIM) is utilized to make a trade-off between the reference and fused images. Experimental results conducted on the synthetic and real-world datasets substantiate the effectiveness and superiority of DRPL compared with other state-of-the-art approaches. The testing code can be found in https://github.com/sasky1/DPRL.
本文提出了一种用于多聚焦图像融合的新型深度网络,名为深度回归对学习(DRPL)。与现有的深度融合方法不同,现有方法将输入图像划分为小补丁并应用分类器来判断补丁是否聚焦,而DRPL直接将整个图像转换为二进制掩码,无需任何补丁操作,随后解决了聚焦/散焦边界周围模糊度估计的难题。同时,一种对学习策略被引入到模型中,该策略将一对互补的源图像作为输入并生成两个相应的二进制掩码,极大地对每一对图像施加了互补约束,并对性能提升做出了巨大贡献。此外,由于聚焦部分存在边缘或梯度,而散焦部分没有类似特性,我们还嵌入了梯度损失以确保生成的图像完全聚焦。然后利用结构相似性指数(SSIM)在参考图像和融合图像之间进行权衡。在合成数据集和真实世界数据集上进行的实验结果证实了DRPL与其他现有最先进方法相比的有效性和优越性。测试代码可在https://github.com/sasky1/DPRL中找到。