Recent years have witnessed significant progress in the field of neural surface reconstruction. While the extensive focus was put on volumetric and implicit approaches, a number of works have shown that explicit graphics primitives such as point clouds can significantly reduce computational complexity, without sacrificing the reconstructed surface quality. However, less emphasis has been put on modeling dynamic surfaces with point primitives. In this work, we present a dynamic point field model that combines the representational benefits of explicit point-based graphics with implicit deformation networks to allow efficient modeling of non-rigid 3D surfaces. Using explicit surface primitives also allows us to easily incorporate well-established constraints such as-isometric-as-possible regularisation. While learning this deformation model is prone to local optima when trained in a fully unsupervised manner, we propose to additionally leverage semantic information such as keypoint dynamics to guide the deformation learning. We demonstrate our model with an example application of creating an expressive animatable human avatar from a collection of 3D scans. Here, previous methods mostly rely on variants of the linear blend skinning paradigm, which fundamentally limits the expressivity of such models when dealing with complex cloth appearances such as long skirts. We show the advantages of our dynamic point field framework in terms of its representational power, learning efficiency, and robustness to out-of-distribution novel poses.
翻译:近年来,神经表面重建领域取得了重要进展。虽然主要集中在体积和隐式方法上,但一些研究表明,明确的图形基元,如点云,可以显著降低计算复杂度,而不牺牲重建表面质量。但是,点基元动态表面建模方面研究较少。在本文中,我们提出了一种动态点场模型,它结合了显式点基图形的表示优势和隐式变形网络,实现了非刚性3D表面的高效建模。使用显式表面基元还使我们轻松地纳入已经建立的约束,例如尽可能等距正则化。虽然完全不受监督的方式学习变形模型容易陷入局部最优,但我们建议另外利用语义信息,例如关键点动态,以指导变形学习。我们通过一个示例展示了模型,该示例应用于从一组3D扫描中创建具有表现力的可动人体化身。在这里,以前的方法大多依赖于线性混合绑定皮肤范例,当处理长裙等复杂布料外观时,这在根本上限制了这些模型的表现力。我们展示了动态点场框架在表示能力、学习效率和对分布外的新颖姿势的稳健性方面的优点。