Generative Adversarial Networks (GANs) are commonly used for modeling complex distributions of data. Both the generators and discriminators of GANs are often modeled by neural networks, posing a non-transparent optimization problem which is non-convex and non-concave over the generator and discriminator, respectively. Such networks are often heuristically optimized with gradient descent-ascent (GDA), but it is unclear whether the optimization problem contains any saddle points, or whether heuristic methods can find them in practice. In this work, we analyze the training of Wasserstein GANs with two-layer neural network discriminators through the lens of convex duality, and for a variety of generators expose the conditions under which Wasserstein GANs can be solved exactly with convex optimization approaches, or can be represented as convex-concave games. Using this convex duality interpretation, we further demonstrate the impact of different activation functions of the discriminator. Our observations are verified with numerical results demonstrating the power of the convex interpretation, with applications in progressive training of convex architectures corresponding to linear generators and quadratic-activation discriminators for CelebA image generation. The code for our experiments is available at https://github.com/ardasahiner/ProCoGAN.
翻译:生成自动网络( GANs) 通常用于模拟复杂的数据分布。 GANs 的生成者和歧视者通常都以神经网络为模型, 形成一个不透明的优化问题, 对生成者和歧视者来说, 前者的生成者与歧视者而言, 后者的生成者与歧视者之间互不相容。 这种网络通常使用渐渐渐的下降率( GDA) 来优化, 但目前还不清楚优化问题是否包含任何支撑点, 或超自然方法在实践中能否找到它们。 在这项工作中, 我们通过对相交双层神经网络歧视者进行模拟, 并分析各种生成者所面临的不透明优化问题, 后者对生成者而言, 后者的优化方法完全可以解决, 或可被描述为渐渐渐的下降率游戏( GDA ) 。 使用这种配置的双重解释, 我们进一步展示了歧视者的不同激活功能的影响 。 我们的观察结果通过数字来验证了对二层神经网络解释的力度, 其应用在相交式神经网络的图像生成者/ CELA Prodasiax 用于Crecialal- condistrax 用于生成的Calal- comfitracial- comcial- comtix comtix 。