Guo Xiaopeng, Nie Rencan, Cao Jinde, Zhou Dongming, Qian Wenhua
School of Information Science and Engineering, Yunnan University, Kunming, Yunnan 650091, China
School of Information Science and Engineering, Yunnan University, Kunming, Yunnan 650091, China, and School of Automation, Southeast University, Jiangsu, Nanjing 210096, China
Neural Comput. 2018 Jul;30(7):1775-1800. doi: 10.1162/neco_a_01098. Epub 2018 Jun 12.
As the optical lenses for cameras always have limited depth of field, the captured images with the same scene are not all in focus. Multifocus image fusion is an efficient technology that can synthesize an all-in-focus image using several partially focused images. Previous methods have accomplished the fusion task in spatial or transform domains. However, fusion rules are always a problem in most methods. In this letter, from the aspect of focus region detection, we propose a novel multifocus image fusion method based on a fully convolutional network (FCN) learned from synthesized multifocus images. The primary novelty of this method is that the pixel-wise focus regions are detected through a learning FCN, and the entire image, not just the image patches, are exploited to train the FCN. First, we synthesize 4500 pairs of multifocus images by repeatedly using a gaussian filter for each image from PASCAL VOC 2012, to train the FCN. After that, a pair of source images is fed into the trained FCN, and two score maps indicating the focus property are generated. Next, an inversed score map is averaged with another score map to produce an aggregative score map, which take full advantage of focus probabilities in two score maps. We implement the fully connected conditional random field (CRF) on the aggregative score map to accomplish and refine a binary decision map for the fusion task. Finally, we exploit the weighted strategy based on the refined decision map to produce the fused image. To demonstrate the performance of the proposed method, we compare its fused results with several start-of-the-art methods not only on a gray data set but also on a color data set. Experimental results show that the proposed method can achieve superior fusion performance in both human visual quality and objective assessment.
由于相机的光学镜头景深总是有限的,所以拍摄的同一场景图像并非都清晰对焦。多聚焦图像融合是一种高效技术,它可以利用多幅部分对焦的图像合成一幅全对焦图像。以往的方法是在空间域或变换域完成融合任务。然而,在大多数方法中,融合规则始终是个问题。在这篇文章中,我们从聚焦区域检测的角度,提出了一种基于从合成多聚焦图像中学习得到的全卷积网络(FCN)的新型多聚焦图像融合方法。该方法的主要新颖之处在于通过学习FCN检测逐像素的聚焦区域,并且利用整个图像(而非仅仅图像块)来训练FCN。首先,我们通过对PASCAL VOC 2012中的每幅图像重复使用高斯滤波器来合成4500对多聚焦图像,用于训练FCN。之后,将一对源图像输入到训练好的FCN中,生成两个表示聚焦特性的得分图。接下来,将一个反转得分图与另一个得分图求平均,以生成一个聚合得分图,该图充分利用了两个得分图中的聚焦概率。我们在聚合得分图上实现全连接条件随机场(CRF),以完成并细化用于融合任务的二值决策图。最后,我们基于细化后的决策图采用加权策略来生成融合图像。为了证明所提方法的性能,我们不仅在灰度数据集上,还在彩色数据集上,将其融合结果与几种最新方法进行了比较。实验结果表明,所提方法在人类视觉质量和客观评估方面均能实现卓越的融合性能。