Generative Adversarial Networks (GANs) are powerful generative models, but suffer from training instability. The recently proposed Wasserstein GAN (WGAN) makes progress toward stable training of GANs, but sometimes can still generate only low-quality samples or fail to converge. We find that these problems are often due to the use of weight clipping in WGAN to enforce a Lipschitz constraint on the critic, which can lead to undesired behavior. We propose an alternative to clipping weights: penalize the norm of gradient of the critic with respect to its input. Our proposed method performs better than standard WGAN and enables stable training of a wide variety of GAN architectures with almost no hyperparameter tuning, including 101-layer ResNets and language models over discrete data. We also achieve high quality generations on CIFAR-10 and LSUN bedrooms.
翻译:我们发现,这些问题往往是由于工作组使用权重剪切法对评论家施加利普施奇茨限制,这可能导致不可取的行为。我们建议了剪切权重的替代办法:惩罚批评者投入的梯度标准。我们提议的方法比标准WGAN要好,能够对几乎没有超光度调整的各类GAN结构进行稳定培训,包括101级ResNet和语言模型,以取代离散数据。我们还在CIFAR-10和LSUN卧室实现了高素质世代。