Training a Neural Radiance Field (NeRF) without pre-computed camera poses is challenging. Recent advances in this direction demonstrate the possibility of jointly optimising a NeRF and camera poses in forward-facing scenes. However, these methods still face difficulties during dramatic camera movement. We tackle this challenging problem by incorporating undistorted monocular depth priors. These priors are generated by correcting scale and shift parameters during training, with which we are then able to constrain the relative poses between consecutive frames. This constraint is achieved using our proposed novel loss functions. Experiments on real-world indoor and outdoor scenes show that our method can handle challenging camera trajectories and outperforms existing methods in terms of novel view rendering quality and pose estimation accuracy.
翻译:培训神经辐射场(NERF)时没有预先配置的相机,这是具有挑战性的。这方面的最近进展表明,有可能联合优化NERF和照相机在远视场中安装。然而,这些方法在镜头突变时仍面临困难。我们通过纳入非扭曲的单眼深度前科来应对这一具有挑战性的问题。这些前科是通过在训练期间纠正规模和改变参数产生的,然后通过这些参数,我们就可以限制连续框架之间的相对结构。这一限制是通过我们拟议的新的损失功能实现的。在现实世界的室内和室外场景实验显示,我们的方法可以处理具有挑战性的相机轨迹,并超越现有方法,用新颖的观点来提供质量和估计准确性。