Yunnan Key Laboratory of Opto-electronic Information Technology, Yunnan Normal University, Kunming, 650500, China.
Department of Thoracic Surgery, The First People's Hospital of Yunnan Province, Kunming, 650500, China.
Int J Comput Assist Radiol Surg. 2024 May;19(5):951-960. doi: 10.1007/s11548-024-03080-8. Epub 2024 Feb 27.
In virtual surgery, the appearance of 3D models constructed from CT images lacks realism, leading to potential misunderstandings among residents. Therefore, it is crucial to reconstruct realistic endoscopic scene using multi-view images captured by an endoscope.
We propose an Endoscope-NeRF network for implicit radiance fields reconstruction of endoscopic scene under non-fixed light source, and synthesize novel views using volume rendering. Endoscope-NeRF network with multiple MLP networks and a ray transformer network represents endoscopic scene as implicit field function with color and volume density at continuous 5D vectors (3D position and 2D direction). The final synthesized image is obtained by aggregating all sampling points on each ray of the target camera using volume rendering. Our method considers the effect of distance from the light source to the sampling point on the scene radiance.
Our network is validated on the lung, liver, kidney and heart of pig collected by our device. The results show that the novel views of endoscopic scene synthesized by our method outperform existing methods (NeRF and IBRNet) in terms of PSNR, SSIM, and LPIPS metrics.
Our network can effectively learn a radiance field function with generalization ability. Fine-tuning the pre-trained model on a new endoscopic scene to further optimize the neural radiance fields of the scene, which can provide more realistic, high-resolution rendered images for surgical simulation.
在虚拟手术中,由 CT 图像构建的 3D 模型外观缺乏真实感,导致住院医师之间存在潜在误解。因此,使用内窥镜捕获的多视图图像重建逼真的内窥镜场景至关重要。
我们提出了一种用于非固定光源下内窥镜场景的隐式辐射场重建的内窥镜-NeRF 网络,并使用体绘制合成新视图。内窥镜-NeRF 网络由多个 MLP 网络和射线变换网络组成,将内窥镜场景表示为连续 5D 向量(3D 位置和 2D 方向)的隐式场函数,具有颜色和体积密度。最终的合成图像是通过在目标相机的每条射线的所有采样点上使用体绘制进行聚合而获得的。我们的方法考虑了场景辐射度从光源到采样点的距离对场景辐射度的影响。
我们的网络在我们的设备采集的猪的肺、肝、肾和心脏上进行了验证。结果表明,与现有方法(NeRF 和 IBRNet)相比,我们的方法合成的内窥镜场景新视图在 PSNR、SSIM 和 LPIPS 指标方面表现更好。
我们的网络可以有效地学习具有泛化能力的辐射场函数。在新的内窥镜场景上微调预训练模型,以进一步优化场景的神经辐射场,从而为手术模拟提供更真实、高分辨率的渲染图像。