The existing auto-encoder based face pose editing methods primarily focus on modeling the identity preserving ability during pose synthesis, but are less able to preserve the image style properly, which refers to the color, brightness, saturation, etc. In this paper, we take advantage of the well-known frontal/profile optical illusion and present a novel two-stage approach to solve the aforementioned dilemma, where the task of face pose manipulation is cast into face inpainting. By selectively sampling pixels from the input face and slightly adjust their relative locations with the proposed ``Pixel Attention Sampling" module, the face editing result faithfully keeps the identity information as well as the image style unchanged. By leveraging high-dimensional embedding at the inpainting stage, finer details are generated. Further, with the 3D facial landmarks as guidance, our method is able to manipulate face pose in three degrees of freedom, i.e., yaw, pitch, and roll, resulting in more flexible face pose editing than merely controlling the yaw angle as usually achieved by the current state-of-the-art. Both the qualitative and quantitative evaluations validate the superiority of the proposed approach.
翻译:现有基于自动编码器面部的编辑方法主要侧重于在合成合成时对身份保存能力进行模型化,但不太能够适当保存图像样式,即颜色、亮度、饱和度等。 在本文件中,我们利用众所周知的正面/显性光学幻觉,提出了解决上述困境的新颖的两阶段办法,即将面部任务置于操纵上,将面部任务置于面部涂色中。通过有选择地从输入面部采样,并用拟议的“Pixel注意取样”模块略微调整其相对位置,面部编辑结果忠实地保持身份信息以及图像样式不变。通过在油漆阶段利用高维嵌入,生成了更精细的细节。此外,以3D面部的标志为指导,我们的方法能够在3度上操纵面部面部姿势,即:yaw、声道和滚动,导致面部面部面部更灵活的面部面部构成,而不是仅仅控制当前状态通常达到的雅角度。定性和定量评价都验证了拟议方法的优越性。