In the literature, 3D reconstruction from 2D image has been extensively addressed but often still requires geometrical supervision. In this paper, we propose SceneRF, a self-supervised monocular scene reconstruction method with neural radiance fields (NeRF) learned from multiple image sequences with pose. To improve geometry prediction, we introduce new geometry constraints and a novel probabilistic sampling strategy that efficiently update radiance fields. As the latter are conditioned on a single frame, scene reconstruction is achieved from the fusion of multiple synthesized novel depth views. This is enabled by our spherical-decoder, which allows hallucination beyond the input frame field of view. Thorough experiments demonstrate that we outperform all baselines on all metrics for novel depth views synthesis and scene reconstruction. Our code is available at https://astra-vision.github.io/SceneRF.
翻译:在文献中,从 2D 图像重建 3D 已得到广泛处理,但往往仍然需要几何监督。 在本文中,我们提议SceenRF, 这是一种自我监督的单视场重建方法,由神经光谱场(NeRF)从多重图像序列中学习。为了改进几何预测,我们引入了新的几何限制和新的概率抽样战略,以有效更新光谱场。由于后者以单一框架为条件,现场重建是通过多个合成新深度视图的融合而实现的。这是由我们的球形分解器促成的,它允许幻觉超出输入框架视野范围。索罗试验表明,我们超越了所有指标的所有基线,用于新的深度观点合成和场重建。我们的代码可以在 https://astra-vision.github.io/SceenRF 上查阅。