Woo Sung-Min, Ryu Je-Ho, Kim Jong-Ok
IEEE Trans Image Process. 2021;30:5001-5016. doi: 10.1109/TIP.2021.3077137. Epub 2021 May 19.
Multi-exposure image fusion inevitably causes ghost artifacts owing to inaccurate image registration. In this study, we propose a deep learning technique for the seamless fusion of multi-exposed low dynamic range (LDR) images using a focus-pixel sensor. For auto-focusing in mobile cameras, a focus-pixel sensor originally provides left (L) and right (R) luminance images simultaneously with a full-resolution RGB image. These L/R images are less saturated than the RGB images because they are summed up to be a normal pixel value in the RGB image of the focus pixel sensor. These two features of the focus pixel image, namely, relatively short exposure and perfect alignment are utilized in this study to provide fusion cues for high dynamic range (HDR) imaging. To minimize fusion artifacts, luminance and chrominance fusions are performed separately in two sub-nets. In a luminance recovery network, two heterogeneous images, the focus pixel image and the corresponding overexposed LDR image, are first fused by joint learning to produce an HDR luminance image. Subsequently, a chrominance network fuses the color components of the misaligned underexposed LDR input to obtain a 3-channel HDR image. Existing deep-neural-network-based HDR fusion methods fuse misaligned multi-exposed inputs directly. They suffer from visual artifacts that are observed mostly in saturated regions because pixel values are clipped out. Meanwhile, the proposed method reconstructs missing luminance with aligned unsaturated focus pixel image first, and thus, the luma-recovered image provides the cues for accurate color fusion. The experimental results show that the proposed method not only accurately restores fine details in saturated areas, but also produce ghost-free high-quality HDR images without pre-alignment.
多曝光图像融合由于图像配准不准确,不可避免地会产生重影伪像。在本研究中,我们提出了一种深度学习技术,用于使用对焦像素传感器对多曝光低动态范围(LDR)图像进行无缝融合。对于移动相机中的自动对焦,对焦像素传感器最初会同时提供左(L)和右(R)亮度图像以及全分辨率RGB图像。这些L/R图像的饱和度低于RGB图像,因为它们在对焦像素传感器的RGB图像中被求和成为一个正常像素值。本研究利用对焦像素图像的这两个特性,即相对较短的曝光和完美对齐,为高动态范围(HDR)成像提供融合线索。为了最小化融合伪像,亮度和色度融合在两个子网中分别进行。在亮度恢复网络中,首先通过联合学习将两个异质图像,即对焦像素图像和相应的过曝光LDR图像进行融合,以生成HDR亮度图像。随后,色度网络融合未对齐的欠曝光LDR输入的颜色分量,以获得3通道HDR图像。现有的基于深度神经网络的HDR融合方法直接融合未对齐的多曝光输入。它们会出现视觉伪像,这些伪像大多出现在饱和区域,因为像素值被裁剪掉了。同时,所提出的方法首先用对齐的不饱和对焦像素图像重建缺失的亮度,因此,亮度恢复后的图像为准确的颜色融合提供了线索。实验结果表明,所提出的方法不仅能准确恢复饱和区域的精细细节,还能在无需预对齐的情况下生成无重影的高质量HDR图像。