Face image manipulation via three-dimensional guidance has been widely applied in various interactive scenarios due to its semantically-meaningful understanding and user-friendly controllability. However, existing 3D-morphable-model-based manipulation methods are not directly applicable to out-of-domain faces, such as non-photorealistic paintings, cartoon portraits, or even animals, mainly due to the formidable difficulties in building the model for each specific face domain. To overcome this challenge, we propose, as far as we know, the first method to manipulate faces in arbitrary domains using human 3DMM. This is achieved through two major steps: 1) disentangled mapping from 3DMM parameters to the latent space embedding of a pre-trained StyleGAN2 that guarantees disentangled and precise controls for each semantic attribute; and 2) cross-domain adaptation that bridges domain discrepancies and makes human 3DMM applicable to out-of-domain faces by enforcing a consistent latent space embedding. Experiments and comparisons demonstrate the superiority of our high-quality semantic manipulation method on a variety of face domains with all major 3D facial attributes controllable-pose, expression, shape, albedo, and illumination. Moreover, we develop an intuitive editing interface to support user-friendly control and instant feedback. Our project page is https://cassiepython.github.io/cddfm3d/index.html
翻译:通过三维指导对脸部进行面部操纵,在各种互动情景中广泛应用了通过三维指导进行的脸部操纵,因为其具有内在意义的理解和用户友好的可控性。然而,现有的3D可变模型操纵方法并不直接适用于外体面部,例如非光真真切的绘画、卡通肖像,甚至动物,主要原因是在为每个特定面域建立模型时存在巨大的困难。为了克服这一挑战,我们建议,根据我们所知,在任意域使用人类3DMMM进行面部操控的第一种方法。这通过两个主要步骤实现:1)从3DMM参数到隐蔽空间嵌入预先训练的StelegGAN2,以保障对每个语义属性的分解和精确控制;以及2)交叉调整,以弥合区域差异并使人类3DMMM适用于外体面部的模型,通过执行一致的隐蔽空间嵌入。实验和比较表明,我们高品质的语界操控方法在多个面域的优越性方法上,所有主要的3D面基3GLD3的页面界面接口接口、我们可以进行直观的缩缩、直观的图像的缩成。