Generative adversarial networks (GANs) are successfully used for image synthesis but are known to face instability during training. In contrast, probabilistic diffusion models (DMs) are stable and generate high-quality images, at the cost of an expensive sampling procedure. In this paper, we introduce a simple method to allow GANs to stably converge to their theoretical optimum, while bringing in the denoising machinery from DMs. These models are combined into a simpler model (ATME) that only requires a forward pass during inference, making predictions cheaper and more accurate than DMs and popular GANs. ATME breaks an information asymmetry existing in most GAN models in which the discriminator has spatial knowledge of where the generator is failing. To restore the information symmetry, the generator is endowed with knowledge of the entropic state of the discriminator, which is leveraged to allow the adversarial game to converge towards equilibrium. We demonstrate the power of our method in several image-to-image translation tasks, showing superior performance than state-of-the-art methods at a lesser cost. Code is available at https://github.com/DLR-MI/atme
翻译:生成对抗网络(GAN)已成功应用于图像合成,但已知在训练过程中面临不稳定性。相反,概率扩散模型(DM)稳定且可生成高质量图像,但需要昂贵的抽样过程。在本文中,我们介绍了一种简单的方法,使GAN能够稳定地收敛到其理论最优解,同时引入DM的去噪机制。这些模型组合成一个更简单的模型(ATME),仅在推理期间需要前向传递,使预测比DM和流行的GAN更便宜、更准确。ATME打破了大多数GAN模型中存在的信息不对称,其中鉴别器具有生成器失败位置的空间知识。为了恢复信息对称性,生成器被赋予鉴别器的熵状态知识,该知识被利用以使对抗游戏收敛到平衡点。我们在几个图像到图像转换任务中展示了我们方法的强大表现,证明了其比最先进的方法更具优越性,成本更低。代码可在 https://github.com/DLR-MI/atme 获得。