3D reconstruction and novel view synthesis of dynamic scenes from collections of single views recently gained increased attention. Existing work shows impressive results for synthetic setups and forward-facing real-world data, but is severely limited in the training speed and angular range for generating novel views. This paper addresses these limitations and proposes a new method for full 360{\deg} novel view synthesis of non-rigidly deforming scenes. At the core of our method are: 1) An efficient deformation module that decouples the processing of spatial and temporal information for acceleration at training and inference time; and 2) A static module representing the canonical scene as a fast hash-encoded neural radiance field. We evaluate the proposed approach on the established synthetic D-NeRF benchmark, that enables efficient reconstruction from a single monocular view per time-frame randomly sampled from a full hemisphere. We refer to this form of inputs as monocularized data. To prove its practicality for real-world scenarios, we recorded twelve challenging sequences with human actors by sampling single frames from a synchronized multi-view rig. In both cases, our method is trained significantly faster than previous methods (minutes instead of days) while achieving higher visual accuracy for generated novel views. Our source code and data is available at our project page https://graphics.tu-bs.de/publications/kappel2022fast.
翻译:现有工作显示合成集成和前瞻性真实世界数据方面令人印象深刻的结果,但培训速度和角范围严重有限,难以产生新观点。本文件述及这些局限性,并提出了一个新的方法,用于对非硬性变形场景进行全360×deg}新视角合成。我们方法的核心是:(1) 一个高效的变形模块,它分解了空间和时间信息的处理,以便在培训和推断时间加快速度;和(2) 一个固定模块,作为快速的集成编码神经光谱场,代表了金星场。我们评估了既定的合成D-NERF基准的拟议方法,从每个时间框架随机抽样抽取的单一单一视角中可以有效地重建整个半球。我们把这种输入形式称为单层化数据。为了证明它对于现实世界情景的实用性,我们记录了12个具有挑战性的序列,从一个同步的多视角的单一框架中取样。在两种情况下,我们的方法都是以新颖的单一视角/直径对我们的源码进行快速的训练。在以往的页/直径上,我们的数据源的精确度上,我们的方法比我们的直观/直径的版本要快。