Neural Radiance Field (NeRF) has recently emerged as a powerful representation to synthesize photorealistic novel views. While showing impressive performance, it relies on the availability of dense input views with highly accurate camera poses, thus limiting its application in real-world scenarios. In this work, we introduce Sparse Pose Adjusting Radiance Field (SPARF), to address the challenge of novel-view synthesis given only few wide-baseline input images (as low as 3) with noisy camera poses. Our approach exploits multi-view geometry constraints in order to jointly learn the NeRF and refine the camera poses. By relying on pixel matches extracted between the input views, our multi-view correspondence objective enforces the optimized scene and camera poses to converge to a global and geometrically accurate solution. Our depth consistency loss further encourages the reconstructed scene to be consistent from any viewpoint. Our approach sets a new state of the art in the sparse-view regime on multiple challenging datasets.
翻译:神经辐射场(NeRF)最近已成为合成逼真新视图的强大表示方法。尽管表现出令人印象深刻的性能,但它依赖于密集输入视图和高精度相机姿势的可用性,从而限制了它在实际场景中的应用。在这项工作中,我们介绍了Sparse Pose Adjusting Radiance Field(SPARF),以解决给定少量宽基线输入图像(尽低至3张)和嘈杂相机姿势的新视图综合挑战。我们的方法利用多视角几何约束,以共同学习神经辐射场并优化相机姿势。通过基于提取自输入视图的像素匹配,我们的多视角对应目标将优化的场景和相机姿势强制收敛为全局和几何精确的解决方案。我们的深度一致性损失进一步鼓励从任何视点观察重建的场景保持一致性。我们的方法在多个具有挑战性的数据集中设立了稀疏视角方面的新纪录。