In this paper, we focus on the challenges of modeling deformable 3D objects from casual videos. With the popularity of neural radiance fields (NeRF), many works extend it to dynamic scenes with a canonical NeRF and a deformation model that achieves 3D point transformation between the observation space and the canonical space. Recent works rely on linear blend skinning (LBS) to achieve the canonical-observation transformation. However, the linearly weighted combination of rigid transformation matrices is not guaranteed to be rigid. As a matter of fact, unexpected scale and shear factors often appear. In practice, using LBS as the deformation model can always lead to skin-collapsing artifacts for bending or twisting motions. To solve this problem, we propose neural dual quaternion blend skinning (NeuDBS) to achieve 3D point deformation, which can perform rigid transformation without skin-collapsing artifacts. Besides, we introduce a texture filtering approach for texture rendering that effectively minimizes the impact of noisy colors outside target deformable objects. Extensive experiments on real and synthetic datasets show that our approach can reconstruct 3D models for humans and animals with better qualitative and quantitative performance than state-of-the-art methods.
翻译:本文关注从非结构化视频中模拟变形三维物体的挑战。随着神经放射场(NeRF)的流行,许多研究将其扩展到具有规范NeRF和变形模型的动态场景,实现观察空间和规范空间之间的三维点变换。最近的研究依赖于线性混合蒙皮(LBS)来实现规范-观测转换。然而,刚性变换矩阵的线性加权组合不能保证刚性。事实上,意外的比例和剪切因子经常出现。在实践中,使用LBS作为变形模型总是会导致弯曲或扭曲运动的皮肤坍塌伪影。为了解决这个问题,我们提出了神经双四元混合蒙皮(NeuDBS)来实现3D点变形,可以在没有皮肤坍塌伪像的情况下执行刚性变换。此外,我们引入了一种纹理过滤方法,用于有效地减少目标可变形物体外部嘈杂色彩的影响。对真实和合成数据集的广泛实验表明,与现有方法相比,我们的方法可以更好地定量和定性地重建人和动物的3D模型。