Recent years have witnessed significant progress in the field of neural surface reconstruction. While the extensive focus was put on volumetric and implicit approaches, a number of works have shown that explicit graphics primitives such as point clouds can significantly reduce computational complexity, without sacrificing the reconstructed surface quality. However, less emphasis has been put on modeling dynamic surfaces with point primitives. In this work, we present a dynamic point field model that combines the representational benefits of explicit point-based graphics with implicit deformation networks to allow efficient modeling of non-rigid 3D surfaces. Using explicit surface primitives also allows us to easily incorporate well-established constraints such as-isometric-as-possible regularisation. While learning this deformation model is prone to local optima when trained in a fully unsupervised manner, we propose to additionally leverage semantic information such as keypoint dynamics to guide the deformation learning. We demonstrate our model with an example application of creating an expressive animatable human avatar from a collection of 3D scans. Here, previous methods mostly rely on variants of the linear blend skinning paradigm, which fundamentally limits the expressivity of such models when dealing with complex cloth appearances such as long skirts. We show the advantages of our dynamic point field framework in terms of its representational power, learning efficiency, and robustness to out-of-distribution novel poses.
翻译:近年来,在神经表面重建领域取得了重大进展。尽管广泛关注体积和隐式方法,但一些工作表明,显式图形原语如点云可以显著减少计算复杂度,而无需牺牲重建表面质量。然而,在使用点原语建模动态表面方面,人们并没有过多地关注。在本文中,我们提出了一个动态点场模型,结合了显式基于点的图形的表征优势和隐式变形网络,以实现非刚性3D表面的有效建模。使用显式表面原语还允许我们轻松地结合享有盛誉的约束,如尽量等距正则化。虽然完全无监督地训练这个变形模型容易陷入局部最优解,但我们建议额外利用语义信息(如关键点动态)来指导变形学习。我们通过以从一系列3D扫描创建一个有表现力的可动态表现的人类化身的示例应用演示了我们的模型。在这里,先前的方法主要依赖于线性混合皮肤范例的变体,这在处理长裙等复杂布料外观时基本上限制了这种模型的表现力。我们展示了我们动态点场框架在其表征能力、学习效率和对于分布外新姿势的鲁棒性方面的优势。