Rendering novel views from captured multi-view images has made considerable progress since the emergence of the neural radiance field. This paper aims to further advance the quality of view rendering by proposing a novel approach dubbed the neural radiance feature field (NRFF) which represents scenes in the feature space. We first propose a multiscale tensor decomposition scheme to organize learnable features so as to represent scenes from coarse to fine scales. We demonstrate many benefits of the proposed multiscale representation, including more accurate scene shape and appearance reconstruction, and faster convergence compared with the single-scale representation. Instead of encoding view directions to model view-dependent effects, we further propose to encode the rendering equation in the feature space by employing the anisotropic spherical Gaussian mixture predicted from the proposed multiscale representation. The proposed NRFF improves state-of-the-art rendering results by over 1 dB in PSNR on both the NeRF and NSVF synthetic datasets. A significant improvement has also been observed on the real-world Tanks and Temples dataset.
翻译:自神经弧度场落以来,从所拍摄的多视图图像中产生的新观点取得了相当大的进展。本文件旨在通过提出一种称为神经亮度特征场(NRFF)的新颖方法来进一步提高视觉质量,该方法代表地貌中的场景。我们首先提出一个多尺度的阵列分解方案,以组织可学习的特征,从而代表从粗糙到细微的场景。我们展示了拟议的多尺度代表制的许多好处,包括更准确的场景形状和外观重建,以及与单一尺度代表制相比更快的趋同。我们进一步提议使用从拟议的多尺度代表制中预测的厌异球形高斯混合体,在地貌空间编码方程式。拟议的阵列改进了PSNRF和NSVF合成数据集中超过1 dB的艺术效果。在现实世界坦克和坦普尔数据集上也观察到了显著的改进。</s>