Existing neural radiance fields (NeRF) methods for large-scale scene modeling require days of training using multiple GPUs, hindering their applications in scenarios with limited computing resources. Despite fast optimization NeRF variants have been proposed based on the explicit dense or hash grid features, their effectivenesses are mainly demonstrated in object-scale scene representation. In this paper, we point out that the low feature resolution in explicit representation is the bottleneck for large-scale unbounded scene representation. To address this problem, we introduce a new and efficient hybrid feature representation for NeRF that fuses the 3D hash-grids and high-resolution 2D dense plane features. Compared with the dense-grid representation, the resolution of a dense 2D plane can be scaled up more efficiently. Based on this hybrid representation, we propose a fast optimization NeRF variant, called GP-NeRF, that achieves better rendering results while maintaining a compact model size. Extensive experiments on multiple large-scale unbounded scene datasets show that our model can converge in 1.5 hours using a single GPU while achieving results comparable to or even better than the existing method that requires about one day's training with 8 GPUs.
翻译:大型场景模型的现有神经光亮场(NERF)方法要求使用多个GPU进行数天的培训,从而妨碍在有限的计算资源条件下应用这些功能。尽管根据明显的密度或散状网格特征提出了快速优化 NERF 变量,但其效力主要表现在物体规模的场景演示中。在本文中,我们指出,清晰代表的低特征分辨率是大型无限制场景演示的瓶颈。为解决这一问题,我们为NERF引入了新的高效混合特征代表器,该模型将3D hash-grids和高分辨率2D密度平面特征结合在一起。与密度网格代表相比,密度2D平面的分辨率可以更高效地扩大。基于这种混合代表器,我们建议快速优化NERF 变量(称为GP-NERF),该变量在保持一个紧凑的模型大小的同时取得更好的效果。关于多个大型无限制场景数据集的大规模实验表明,我们的模型可以在1.5小时内结合一个GPU,同时取得与大约一天的GPU相比或更好的结果。</s>