We present the first method capable of photorealistically reconstructing deformable scenes using photos/videos captured casually from mobile phones. Our approach augments neural radiance fields (NeRF) by optimizing an additional continuous volumetric deformation field that warps each observed point into a canonical 5D NeRF. We observe that these NeRF-like deformation fields are prone to local minima, and propose a coarse-to-fine optimization method for coordinate-based models that allows for more robust optimization. By adapting principles from geometry processing and physical simulation to NeRF-like models, we propose an elastic regularization of the deformation field that further improves robustness. We show that our method can turn casually captured selfie photos/videos into deformable NeRF models that allow for photorealistic renderings of the subject from arbitrary viewpoints, which we dub "nerfies." We evaluate our method by collecting time-synchronized data using a rig with two mobile phones, yielding train/validation images of the same pose at different viewpoints. We show that our method faithfully reconstructs non-rigidly deforming scenes and reproduces unseen views with high fidelity.
翻译:我们展示了第一个能够利用从移动电话中随意捕捉到的照片/视频对变形场进行光化重建的方法。我们的方法通过优化一个额外的连续体积变异场来增强神经光场(NERF ), 使每个观测到的点都变成一个5D NERF。我们观察到,这些类似 NERF 的变形场很容易被本地迷你, 并提出了一种协调模型的粗到软优化方法, 以便实现更强大的优化。 我们通过将几何处理和物理模拟等像 NERF 一样的模型中的原则修改为变形场, 我们建议对变形场进行弹性调整, 以进一步提高强性。 我们显示, 我们的方法可以将随机的自拍自拍照片/ 影像转化为可变形的 NERF 模型, 从任意的角度来对主题进行摄影真实化的图像转换, 也就是“ 神经元” 。 我们用两部移动电话收集时间同步化数据来评估我们的方法。 我们用不同角度的火车/校准图像, 显示我们的方法忠实地重建了不精确的图像, 。