Facial attribute editing has mainly two objectives: 1) translating image from a source domain to a target one, and 2) only changing the facial regions related to a target attribute and preserving the attribute-excluding details. In this work, we propose a Multi-attention U-Net-based Generative Adversarial Network (MU-GAN). First, we replace a classic convolutional encoder-decoder with a symmetric U-Net-like structure in a generator, and then apply an additive attention mechanism to build attention-based U-Net connections for adaptively transferring encoder representations to complement a decoder with attribute-excluding detail and enhance attribute editing ability. Second, a self-attention mechanism is incorporated into convolutional layers for modeling long-range and multi-level dependencies across image regions. experimental results indicate that our method is capable of balancing attribute editing ability and details preservation ability, and can decouple the correlation among attributes. It outperforms the state-of-the-art methods in terms of attribute manipulation accuracy and image quality.
翻译:外貌属性编辑主要有两个目标:(1) 将图像从源域转换成目标领域;(2) 仅改变与目标属性有关的面部区域,并保存排除属性的细节;在这项工作中,我们提议建立一个多注意的基于U-Net的基因反变异网络(MU-GAN),首先,我们用一个发电机中的对称U-Net结构来取代典型的变相编码解码器,然后应用一个添加式注意机制,建立基于注意的U-Net连接,用于适应性传输编码器表示,以补充带有属性排除细节的解码器,并加强属性编辑能力。第二,将自我注意机制纳入图象区域远距离和多层次依赖性模型的进化层中。实验结果表明,我们的方法能够平衡属性编辑能力和细节保护能力,并能分解属性之间的关联。它超越了属性操纵准确性和图像质量方面的最新方法。