Non-parametric face modeling aims to reconstruct 3D face only from images without shape assumptions. While plausible facial details are predicted, the models tend to over-depend on local color appearance and suffer from ambiguous noise. To address such problem, this paper presents a novel Learning to Aggregate and Personalize (LAP) framework for unsupervised robust 3D face modeling. Instead of using controlled environment, the proposed method implicitly disentangles ID-consistent and scene-specific face from unconstrained photo set. Specifically, to learn ID-consistent face, LAP adaptively aggregates intrinsic face factors of an identity based on a novel curriculum learning approach with relaxed consistency loss. To adapt the face for a personalized scene, we propose a novel attribute-refining network to modify ID-consistent face with target attribute and details. Based on the proposed method, we make unsupervised 3D face modeling benefit from meaningful image facial structure and possibly higher resolutions. Extensive experiments on benchmarks show LAP recovers superior or competitive face shape and texture, compared with state-of-the-art (SOTA) methods with or without prior and supervision.
翻译:虽然预测了可信的面部细节,但模型往往过分依赖本地的颜色外观,并受到模糊的噪音的影响。为了解决这一问题,本文件提出了一个新的“综合和个性化学习”框架(LAP),用于不受监督的强健的3D面部建模。拟议方法不是使用受控制的环境,而是隐含地分解了未受控制的照片组装的ID一致性和场景特有面部。具体地说,为了学习ID一致性的面部,LAP以适应性方式综合了基于新颖课程学习方法的认同的内在面部要素,并放松一致性损失。为了适应个性化场景,我们提议建立一个新型的属性调整网络,用目标属性和细节来修改身份一致的面部。根据拟议方法,我们从有意义的图像面部结构和可能更高的分辨率中获得不受监督的3D面部建模。关于基准的广泛实验显示,LAP在与先期或无监督的状态(SOTA)方法相比,其面部和文字的形状和形状都优或竞争性。