Theoretical analyses for Generative Adversarial Networks (GANs) generally assume an arbitrarily large family of discriminators and do not consider the characteristics of the architectures used in practice. We show that this framework of analysis is too simplistic to properly analyze GAN training. To tackle this issue, we leverage the theory of infinite-width neural networks to model neural discriminator training for a wide range of adversarial losses via its Neural Tangent Kernel (NTK). Our analytical results show that GAN trainability primarily depends on the discriminator's architecture. We further study the discriminator for specific architectures and losses, and highlight properties providing a new understanding of GAN training. For example, we find that GANs trained with the integral probability metric loss minimize the maximum mean discrepancy with the NTK as kernel. Our conclusions demonstrate the analysis opportunities provided by the proposed framework, which paves the way for better and more principled GAN models. We release a generic GAN analysis toolkit based on our framework that supports the empirical part of our study.
翻译:分析结果显示,GAN的可训练性主要取决于歧视者的架构。我们进一步研究了特定架构和损失的制导者,并突出介绍了为GAN培训提供新理解的特性。例如,我们发现,经过培训的GAN以整体概率指数损失培训的特性最大限度地缩小了与NTK作为核心的最小值差异。我们的结论展示了拟议框架提供的分析机会,为更好和更有原则的GAN模式铺平了道路。我们根据支持我们研究经验部分的框架发布了一个通用的GAN分析工具包。