State-of-the-art methods in image-to-image translation are capable of learning a mapping from a source domain to a target domain with unpaired image data. Though the existing methods have achieved promising results, they still produce visual artifacts, being able to translate low-level information but not high-level semantics of input images. One possible reason is that generators do not have the ability to perceive the most discriminative parts between the source and target domains, thus making the generated images low quality. In this paper, we propose a new Attention-Guided Generative Adversarial Networks (AttentionGAN) for the unpaired image-to-image translation task. AttentionGAN can identify the most discriminative foreground objects and minimize the change of the background. The attention-guided generators in AttentionGAN are able to produce attention masks, and then fuse the generation output with the attention masks to obtain high-quality target images. Accordingly, we also design a novel attention-guided discriminator which only considers attended regions. Extensive experiments are conducted on several generative tasks with eight public datasets, demonstrating that the proposed method is effective to generate sharper and more realistic images compared with existing competitive models. The code is available at https://github.com/Ha0Tang/AttentionGAN.
翻译:图像到图像翻译中最先进的方法能够学习从源域到目标域的绘图,使用未受重视的图像数据。虽然现有方法已经取得了令人乐观的成果,但它们仍然产生视觉文物,能够翻译低层次信息,而不是输入图像的高层次语义。一个可能的原因是,生成者没有能力看到源域和目标域之间最具歧视性的部分,从而使生成的图像质量低。在本文中,我们提议为未受重视的图像到图像翻译任务建立一个新的“引人注意的基因分析网络(AttentionGAN) 。“注意GAN”可以识别最具歧视性的地面对象,并最大限度地减少背景变化。“注意GAN”中受关注的生成者能够产生关注面罩,然后将生成输出与关注面罩连接,以获得高质量的目标图像。因此,我们还设计了一个新的关注制导分析器,仅考虑所访问的区域。在与8个公众图像到映的基因化任务上进行了广泛的实验,与8个现实的图像系统进行了对比。“GAN” 演示了现有系统/网络的智能模型是有效的方法。