In this paper, we present a novel double diffusion based neural radiance field, dubbed DD-NeRF, to reconstruct human body geometry and render the human body appearance in novel views from a sparse set of images. We first propose a double diffusion mechanism to achieve expressive representations of input images by fully exploiting human body priors and image appearance details at two levels. At the coarse level, we first model the coarse human body poses and shapes via an unclothed 3D deformable vertex model as guidance. At the fine level, we present a multi-view sampling network to capture subtle geometric deformations and image detailed appearances, such as clothing and hair, from multiple input views. Considering the sparsity of the two level features, we diffuse them into feature volumes in the canonical space to construct neural radiance fields. Then, we present a signed distance function (SDF) regression network to construct body surfaces from the diffused features. Thanks to our double diffused representations, our method can even synthesize novel views of unseen subjects. Experiments on various datasets demonstrate that our approach outperforms the state-of-the-art in both geometric reconstruction and novel view synthesis.
翻译:在本文中,我们展示了一个新颖的双向扩散基于神经光亮场,称为DD-NERF,以重建人体的几何结构,并使人体在一组稀有图像的新观点中出现。我们首先提出一个双向扩散机制,通过充分利用人体的前身和图像外观细节在两个层面实现输入图像的表达式。在粗糙的层次上,我们首先通过一个不穿衣的3D变形脊椎模型来模拟粗体构成和形状,作为指导。在精细层次上,我们提出一个多视角抽样网络,从多个输入视图中捕捉微妙的几何形变形和图像的详细外观,例如服装和毛发。考虑到两个层面特征的宽度,我们将其扩散到能量空间的特性中,以构建神经光亮场。然后,我们展示了一个签名的距离函数(SDFDF)回归网络,从分散的特征中构建人体表面。由于我们的双向分散的表象,我们的方法甚至可以综合对看不见的主体的新观点。对各种数据集进行了实验,在各种数据集图象上显示我们的方法都超越了地球合成的合成。