The goal of this work is to perform 3D reconstruction and novel view synthesis from data captured by scanning platforms commonly deployed for world mapping in urban outdoor environments (e.g., Street View). Given a sequence of posed RGB images and lidar sweeps acquired by cameras and scanners moving through an outdoor scene, we produce a model from which 3D surfaces can be extracted and novel RGB images can be synthesized. Our approach extends Neural Radiance Fields, which has been demonstrated to synthesize realistic novel images for small scenes in controlled settings, with new methods for leveraging asynchronously captured lidar data, for addressing exposure variation between captured images, and for leveraging predicted image segmentations to supervise densities on rays pointing at the sky. Each of these three extensions provides significant performance improvements in experiments on Street View data. Our system produces state-of-the-art 3D surface reconstructions and synthesizes higher quality novel views in comparison to both traditional methods (e.g.~COLMAP) and recent neural representations (e.g.~Mip-NeRF).
翻译:这项工作的目标是进行3D重建和从通常用于在城市室外环境中进行世界制图的扫描平台(例如,Street View)所收集的数据中进行3D重建和新视角合成。鉴于摄影机和扫描机在室外现场移动时获得的成像RGB图像和激光扫瞄系列,我们制作了一个模型,从中可以提取3D表面,并可以合成新型RGB图像。我们的方法扩展了神经光谱场,已经展示了这些场景为受控环境中的小场景综合现实的新图像,并采用了新的方法来利用无同步捕获的Lidar数据,解决所捕捉到的图像之间的暴露变化,并利用预测的图像分割来监督对天空指示的射线的密度。这三个扩展中的每一个都大大改进了对街道视图数据实验的性能。我们的系统制作了最先进的3D表面重建,并合成了与传统方法(例如:~COLMAP)和最近的神经表现(例如~Mip-NERF)相比质量更高的新观点。