Dressed people reconstruction from images is a popular task with promising applications in the creative media and game industry. However, most existing methods reconstruct the human body and garments as a whole with the supervision of 3D models, which hinders the downstream interaction tasks and requires hard-to-obtain data. To address these issues, we propose an unsupervised separated 3D garments and human reconstruction model (USR), which reconstructs the human body and authentic textured clothes in layers without 3D models. More specifically, our method proposes a generalized surface-aware neural radiance field to learn the mapping between sparse multi-view images and geometries of the dressed people. Based on the full geometry, we introduce a Semantic and Confidence Guided Separation strategy (SCGS) to detect, segment, and reconstruct the clothes layer, leveraging the consistency between 2D semantic and 3D geometry. Moreover, we propose a Geometry Fine-tune Module to smooth edges. Extensive experiments on our dataset show that comparing with state-of-the-art methods, USR achieves improvements on both geometry and appearance reconstruction while supporting generalizing to unseen people in real time. Besides, we also introduce SMPL-D model to show the benefit of the separated modeling of clothes and the human body that allows swapping clothes and virtual try-on.
翻译:在创造性媒体和游戏行业中,人们从图像中改制,是一项广受欢迎的任务,在创造性媒体和游戏行业中应用有希望。然而,大多数现有方法在3D模型的监督下,将人体和服装作为一个整体进行重建,这阻碍了下游互动任务,需要难以获取的数据。为了解决这些问题,我们提议了一个不受监督的3D服装和人重建模型(USR),在没有3D模型的情况下,将人体和正正正正正正正正正正正正正正正的纹理服装放在一层中重建。更具体地说,我们的方法提出一个普遍的地表观神经光亮度场,以学习穿衣者稀薄多视图像和地理特征之间的绘图。基于完整的几何测量,我们引入了一种语义和信任引导隔离的隔离战略(SCGS),以探测、分割和重塑服装层。此外,我们提出了一种平滑的地测量精度精度微模模模模模模模模模。关于我们数据设置的广泛实验显示,与州制方法相比,USR在地貌和外观的模型和外观地貌结构重建之间都能够将人类的模型和虚拟模型转换成一面图。