Face editing represents a popular research topic within the computer vision and image processing communities. While significant progress has been made recently in this area, existing solutions: (i) are still largely focused on low-resolution images, (ii) often generate editing results with visual artefacts, or (iii) lack fine-grained control and alter multiple (entangled) attributes at once, when trying to generate the desired facial semantics. In this paper, we aim to address these issues though a novel attribute editing approach called MaskFaceGAN that focuses on local attribute editing. The proposed approach is based on an optimization procedure that directly optimizes the latent code of a pre-trained (state-of-the-art) Generative Adversarial Network (i.e., StyleGAN2) with respect to several constraints that ensure: (i) preservation of relevant image content, (ii) generation of the targeted facial attributes, and (iii) spatially--selective treatment of local image areas. The constraints are enforced with the help of an (differentiable) attribute classifier and face parser that provide the necessary reference information for the optimization procedure. MaskFaceGAN is evaluated in extensive experiments on the CelebA-HQ, Helen and SiblingsDB-HQf datasets and in comparison with several state-of-the-art techniques from the literature, i.e., StarGAN, AttGAN, STGAN, and two versions of InterFaceGAN. Our experimental results show that the proposed approach is able to edit face images with respect to several local facial attributes with unprecedented image quality and at high-resolutions (1024x1024), while exhibiting considerably less problems with attribute entanglement than competing solutions. The source code is made freely available from: https://github.com/MartinPernus/MaskFaceGAN.
翻译:面部编辑在计算机视觉和图像处理社区中是一个流行的研究主题。 虽然最近在这一领域取得了显著进展, 但现有的解决方案有:(一) 仍然主要侧重于低分辨率图像, (二) 经常以视觉手工艺品生成编辑结果, 或 (三) 当试图生成所需的面部语义时, 缺乏精细的控制和同时改变多个( 缠绕的) 属性。 在本文中, 我们的目标是通过一种名为 Mask FaceGAN 的新属性编辑方法来解决这些问题, 重点是本地属性编辑。 拟议的方法基于一种优化程序, 直接优化预培训( 状态- 水平- 艺术) 的潜在代码;(二) 以视觉手动手动手动的图像编辑结果( 即StelegGAN2), 在确保:(一) 保存相关图像内容, (二) 生成有针对性的面部面部特征, (三) 对本地图像区域进行空间选择处理。 限制通过一种( 可理解的) 直径( ) 直径) 和面图像解解码( ) 和面面像( ) 提供为iestal- i- i- i- real- ib- Q) istral- g- 和Silal- gAN 和Silal- dal- dolal- disal) 的文献提供所需的必要参考文档所需的必要引用信息信息信息, 在Syal- disal- disal- 上, 上, 和Syal- disal- disal- disal 上, 和Syal- dal- gal- 和Seal- dal- disal- dal- sal- dal- sal- dal- sal- sal- sal- sal- sal- sal- dal- sal- dal- sal- dalal- dal- sal- disaldald- dal- dal- gal- gal- 和 sal- dal- dal- dal- dal- galdal- dal- gal- gald- dal- dal- dal- dal- dal- gal- dal- dal- dal