Facial expression editing has attracted increasing attention with the advance of deep neural networks in recent years. However, most existing methods suffer from compromised editing fidelity and limited usability as they either ignore pose variations (unrealistic editing) or require paired training data (not easy to collect) for pose controls. This paper presents POCE, an innovative pose-controllable expression editing network that can generate realistic facial expressions and head poses simultaneously with just unpaired training images. POCE achieves the more accessible and realistic pose-controllable expression editing by mapping face images into UV space, where facial expressions and head poses can be disentangled and edited separately. POCE has two novel designs. The first is self-supervised UV completion that allows to complete UV maps sampled under different head poses, which often suffer from self-occlusions and missing facial texture. The second is weakly-supervised UV editing that allows to generate new facial expressions with minimal modification of facial identity, where the synthesized expression could be controlled by either an expression label or directly transplanted from a reference UV map via feature transfer. Extensive experiments show that POCE can learn from unpaired face images effectively, and the learned model can generate realistic and high-fidelity facial expressions under various new poses.
翻译:面部表情编辑随着深度神经网络的进展近年来越来越受到关注。然而,大多数现有的方法要么忽略了姿态变化(编辑效果不真实),要么需要配对的训练数据(难以收集)来控制姿态。本文提出了一种新颖的姿态可控表情编辑网络POCE,只需使用未配对的训练图像就可以同时生成逼真的面部表情和头部姿态。通过将面部图像映射到UV空间,POCE可以实现更易用和更真实的姿态可控表情编辑,从而可以将面部表情和头部姿态分离和单独编辑。POCE具有两个新颖的设计。第一个是自监督的UV完成,可以完成采样在不同头部姿态下的UV映射,这些UV映射经常受到自遮挡和缺失面部纹理的影响。第二个是弱监督的UV编辑,可以通过最小修改面部身份信息生成不同的面部表情,合成的表情可以通过表情标签或通过特征转移从参考UV映射直接移植来控制。大量的实验表明,POCE可以有效地学习未配对的面部图像,并在各种新的姿态下生成逼真且高度保真的面部表情。