Obtaining high-quality 3D reconstructions of room-scale scenes is of paramount importance for upcoming applications in AR or VR. These range from mixed reality applications for teleconferencing, virtual measuring, virtual room planing, to robotic applications. While current volume-based view synthesis methods that use neural radiance fields (NeRFs) show promising results in reproducing the appearance of an object or scene, they do not reconstruct an actual surface. The volumetric representation of the surface based on densities leads to artifacts when a surface is extracted using Marching Cubes, since during optimization, densities are accumulated along the ray and are not used at a single sample point in isolation. Instead of this volumetric representation of the surface, we propose to represent the surface using an implicit function (truncated signed distance function). We show how to incorporate this representation in the NeRF framework, and extend it to use depth measurements from a commodity RGB-D sensor, such as a Kinect. In addition, we propose a pose and camera refinement technique which improves the overall reconstruction quality. In contrast to concurrent work on integrating depth priors in NeRF which concentrates on novel view synthesis, our approach is able to reconstruct high-quality, metrical 3D reconstructions.
翻译:获得高质量的3D对室内场景的重建对于AR或VR的即将应用至关重要。 这些应用包括:电话会议、虚拟测量、虚拟室规划、虚拟室规划、机器人应用等混合现实应用、虚拟计量、虚拟室规划、机器人应用等。目前使用神经弧度场(NERFs)的量基合成方法显示复制物体或场景的有希望的结果,但它们并不重建实际表面。以密度为基础的表面的体积表示在利用三进制立方体提取表面时,会导致文物。此外,在优化期间,密度在射线上积累,而不是单独在一个取样点使用。我们提议使用隐含的功能(经授权的远程功能)代表表面。我们表明如何将这种表示方式纳入NERF框架,并将它扩大到从商品RGB-D传感器(如Kinect)进行深度测量。此外,我们提议了一种表面和摄像改进技术,以提高总体重建质量。与在NRF系统进行整合前的深度工作相比,我们提议在NRF系统重建前的深度方面采用高质量的方法。