Wang Lin, Kim Tae-Kyun, Yoon Kuk-Jin
IEEE Trans Pattern Anal Mach Intell. 2022 Nov;44(11):7657-7673. doi: 10.1109/TPAMI.2021.3113352. Epub 2022 Oct 4.
Event cameras sense brightness changes in each pixel and yield asynchronous event streams instead of producing intensity images. They have distinct advantages over conventional cameras, such as a high dynamic range (HDR) and no motion blur. To take advantage of event cameras with existing image-based algorithms, a few methods have been proposed to reconstruct images from event streams. However, the output images have a low resolution (LR) and are unrealistic. Low-quality outputs stem from broader applications of event cameras, where high-quality and high-resolution (HR) images are needed. In this work, we consider the problem of reconstructing and super-resolving images from LR events when no ground truth (GT) HR images and degradation models are available. We propose a novel end-to-end joint framework for single image reconstruction and super-resolution from LR event data. Our method is primarily unsupervised to handle the absence of real inputs from GT and deploys adversarial learning. To train our framework, we constructed an open dataset, including simulated events and real-world images. The use of the dataset boosts the network performance, and the network architectures and various loss functions in each phase help improve the quality of the resulting image. Various experiments showed that our method surpasses the state-of-the-art LR image reconstruction methods for real-world and synthetic datasets. The experiments for super-resolution (SR) image reconstruction also substantiate the effectiveness of the proposed method. We further extended our method to more challenging problems of HDR, sharp image reconstruction, and color events. In addition, we demonstrate that the reconstruction and super-resolution results serve as intermediate representations of events for high-level tasks, such as semantic segmentation, object recognition, and detection. We further examined how events affect the outputs of the three phases and analyze our method's efficacy through an ablation study.
事件相机感知每个像素的亮度变化并生成异步事件流,而不是生成强度图像。与传统相机相比,它们具有明显的优势,例如高动态范围(HDR)和无运动模糊。为了将事件相机与现有的基于图像的算法结合使用,已经提出了一些从事件流重建图像的方法。然而,输出图像的分辨率较低(LR)且不真实。低质量的输出源于事件相机更广泛的应用场景,在这些场景中需要高质量和高分辨率(HR)的图像。在这项工作中,我们考虑在没有地面真值(GT)HR图像和退化模型的情况下,从低分辨率事件重建和超分辨率图像的问题。我们提出了一种新颖的端到端联合框架,用于从低分辨率事件数据进行单图像重建和超分辨率。我们的方法主要是无监督的,以处理缺乏来自GT的真实输入的情况,并采用对抗学习。为了训练我们的框架,我们构建了一个开放数据集,包括模拟事件和真实世界图像。该数据集的使用提高了网络性能,并且每个阶段的网络架构和各种损失函数有助于提高所得图像的质量。各种实验表明,我们的方法在真实世界和合成数据集上优于现有的低分辨率图像重建方法。超分辨率(SR)图像重建实验也证实了所提出方法的有效性。我们进一步将我们的方法扩展到更具挑战性的高动态范围、清晰图像重建和彩色事件问题。此外,我们证明了重建和超分辨率结果可作为高级任务(如语义分割、目标识别和检测)的事件中间表示。我们进一步研究了事件如何影响三个阶段的输出,并通过消融研究分析了我们方法的有效性。