Lo I-Chan, Chen Homer H
IEEE Trans Image Process. 2023;32:4677-4688. doi: 10.1109/TIP.2023.3299444. Epub 2023 Aug 16.
In this paper, we propose an efficient deep learning pipeline for light field acquisition using a back-to-back dual-fisheye camera. The proposed pipeline generates a light field from a sequence of 360° raw images captured by the dual-fisheye camera. It has three main components: a convolutional network (CNN) that enforces a spatiotemporal consistency constraint on the subviews of the 360° light field, an equirectangular matching cost that aims at increasing the accuracy of disparity estimation, and a light field resampling subnet that produces the 360° light field based on the disparity information. Ablation tests are conducted to analyze the performance of the proposed pipeline using the HCI light field datasets with five objective assessment metrics (MSE, MAE, PSNR, SSIM, and GMSD). We also use real data obtained from a commercially available dual-fisheye camera to quantitatively and qualitatively test the effectiveness, robustness, and quality of the proposed pipeline. Our contributions include: 1) a novel spatiotemporal consistency loss that enforces the subviews of the 360° light field to be consistent, 2) an equirectangular matching cost that combats severe projection distortion of fisheye images, and 3) a light field resampling subnet that retains the geometric structure of spherical subviews while enhancing the angular resolution of the light field.
在本文中,我们提出了一种使用背靠背双鱼眼相机进行光场采集的高效深度学习管道。所提出的管道从双鱼眼相机捕获的360°原始图像序列生成光场。它有三个主要组件:一个对360°光场的子视图施加时空一致性约束的卷积网络(CNN)、一个旨在提高视差估计准确性的等距矩形匹配代价,以及一个基于视差信息生成360°光场的光场重采样子网。使用具有五个客观评估指标(MSE、MAE、PSNR、SSIM和GMSD)的HCI光场数据集进行消融测试,以分析所提出管道的性能。我们还使用从市售双鱼眼相机获得的真实数据,对所提出管道的有效性、鲁棒性和质量进行定量和定性测试。我们的贡献包括:1)一种新颖的时空一致性损失,它使360°光场的子视图保持一致;2)一种应对鱼眼图像严重投影失真的等距矩形匹配代价;3)一个在增强光场角分辨率的同时保留球面子视图几何结构的光场重采样子网。