We present a framework, called MVG-NeRF, that combines classical Multi-View Geometry algorithms and Neural Radiance Fields (NeRF) for image-based 3D reconstruction. NeRF has revolutionized the field of implicit 3D representations, mainly due to a differentiable volumetric rendering formulation that enables high-quality and geometry-aware novel view synthesis. However, the underlying geometry of the scene is not explicitly constrained during training, thus leading to noisy and incorrect results when extracting a mesh with marching cubes. To this end, we propose to leverage pixelwise depths and normals from a classical 3D reconstruction pipeline as geometric priors to guide NeRF optimization. Such priors are used as pseudo-ground truth during training in order to improve the quality of the estimated underlying surface. Moreover, each pixel is weighted by a confidence value based on the forward-backward reprojection error for additional robustness. Experimental results on real-world data demonstrate the effectiveness of this approach in obtaining clean 3D meshes from images, while maintaining competitive performances in novel view synthesis.
翻译:我们提出了一个称为MVG-NERF的框架,将传统的多视几何算法和神经辐射场(NeRF)结合到基于图像的3D重建中。NeRF已经将隐含的3D表层领域革命化,这主要是由于一种不同的量化配方,使高质量的和几何能的新观点合成得以实现。然而,在培训期间,场景的基本几何学没有受到明确的限制,因此在提取一个与行进立方体相交的网块时会导致噪音和不正确的结果。为此,我们提议利用经典的3D重建管道的像素深度和正常度,作为指导 NERF优化的几何前奏。在培训期间,这些前奏被用作假的地面真相,以提高估计基本表面的质量。此外,根据前向后再投射误差,每个像素都经过信任值的权衡,以获得更多的坚固性。关于真实世界数据的实验结果表明这一方法在从图像中获取清洁的3D模束方面是有效的,同时保持新观点合成中的竞争性性表现。