We present Non-Rigid Neural Radiance Fields (NR-NeRF), a reconstruction and novel view synthesis approach for general non-rigid dynamic scenes. Our approach takes RGB images of a dynamic scene as input (e.g., from a monocular video recording), and creates a high-quality space-time geometry and appearance representation. We show that a single handheld consumer-grade camera is sufficient to synthesize sophisticated renderings of a dynamic scene from novel virtual camera views, e.g. a `bullet-time' video effect. NR-NeRF disentangles the dynamic scene into a canonical volume and its deformation. Scene deformation is implemented as ray bending, where straight rays are deformed non-rigidly. We also propose a novel rigidity network to better constrain rigid regions of the scene, leading to more stable results. The ray bending and rigidity network are trained without explicit supervision. Our formulation enables dense correspondence estimation across views and time, and compelling video editing applications such as motion exaggeration. Our code will be open sourced.
翻译:我们为一般非硬性动态场景推出非硬性神经辐射场景(NR-NERRF)的重建和新视角合成方法。我们的方法是将动态场景的RGB图像作为输入(例如单镜录像),并创建高质量的时时时几何和外观代表。我们显示,单手持的消费者级相机足以合成新颖虚拟摄影机视图的动态场景的复杂图像,例如“机床时间”视频效果。NR-NERF将动态场景分解成一个罐头体体积及其变形。屏幕变形作为射线弯曲实施,直射线射线不易变形。我们还提出一个新的僵硬性网络,以更好地限制场景的僵硬区域,导致更稳定的结果。光线弯曲和僵硬性网络在没有明确监督的情况下接受训练。我们的配方能够对各种视图和时间进行密集的通信估计,并强制视频编辑应用,例如运动变形。我们的代码将开源。