Bai Haiyang, Lu Tao, Zhu Jiaqi, Huang Wei, Gou Chang, Guo Jie, Chen Lijun, Guo Yanwen
IEEE Trans Vis Comput Graph. 2025 Oct;31(10):8503-8518. doi: 10.1109/TVCG.2025.3572015.
Recent advancements in neural rendering methods, such as Neural Radiance Fields (NeRF) and 3D Gaussian Splatting (3D-GS), have significantly revolutionized photo-realistic novel view synthesis of scenes with multiple photos or videos as input. However, existing approaches within the NeRF and 3D-GS frameworks often assume the independence of point sampling and ray casting, which are intrinsic to volume rendering and alpha-blending techniques. These underlying assumptions limit the ability to aggregate context within subspaces, such as densities and colors in the radiance fields and pixels on the image plane, leading to synthesized images that lack fine details and smoothness. To overcome this, we propose a universal framework, MixRF, comprising a Radiance Field Mixer (RF-mixer) and a Color Domain Mixer (CD-mixer), to sufficiently aggregate and fully explore information in neighboring sampled points and casting rays, separately. The RF-mixer treats sampled points as an explicit point cloud, enabling the aggregation of density and color attributes from neighboring points to better capture local geometry and appearance. Meanwhile, the CD-mixer rearranges rendered pixels on the sub-image plane, improving smoothness and recovering fine details and textures. Both mixers employ a kernel-based mixing strategy to facilitate effective and controllable attribute aggregation, ensuring a more comprehensive exploration of radiance values and pixel information. Extensive experiments demonstrate that our MixRF framework is compatible with radiance field-based methods, including NeRF and 3D-GS designs. The proposed framework dramatically enhances performance in both qualitative and quantitative evaluations, with less than a $ 25%$25% increase in computational overhead during inference.
神经渲染方法的最新进展,如神经辐射场(NeRF)和3D高斯点云(3D-GS),已经显著革新了以多张照片或视频为输入的场景的逼真新颖视图合成。然而,NeRF和3D-GS框架内的现有方法通常假设点采样和光线投射是独立的,这是体渲染和alpha混合技术所固有的。这些潜在假设限制了在子空间内聚合上下文的能力,例如辐射场中的密度和颜色以及图像平面上的像素,导致合成图像缺乏精细细节和平滑度。为了克服这一问题,我们提出了一个通用框架MixRF,它由一个辐射场混合器(RF混合器)和一个颜色域混合器(CD混合器)组成,以分别充分聚合和全面探索相邻采样点和投射光线中的信息。RF混合器将采样点视为一个显式点云,能够聚合相邻点的密度和颜色属性,以更好地捕捉局部几何形状和外观。同时,CD混合器在子图像平面上重新排列渲染的像素,提高平滑度并恢复精细细节和纹理。两个混合器都采用基于内核的混合策略来促进有效且可控的属性聚合,确保更全面地探索辐射值和像素信息。大量实验表明,我们的MixRF框架与基于辐射场的方法兼容,包括NeRF和3D-GS设计。所提出的框架在定性和定量评估中都显著提高了性能,推理过程中的计算开销增加不到25%。