The instability of Generative Adversarial Network (GAN) training has frequently been attributed to gradient descent. Consequently, recent methods have aimed to tailor the models and training procedures to stabilise the discrete updates. In contrast, we study the continuous-time dynamics induced by GAN training. Both theory and toy experiments suggest that these dynamics are in fact surprisingly stable. From this perspective, we hypothesise that instabilities in training GANs arise from the integration error in discretising the continuous dynamics. We experimentally verify that well-known ODE solvers (such as Runge-Kutta) can stabilise training - when combined with a regulariser that controls the integration error. Our approach represents a radical departure from previous methods which typically use adaptive optimisation and stabilisation techniques that constrain the functional space (e.g. Spectral Normalisation). Evaluation on CIFAR-10 and ImageNet shows that our method outperforms several strong baselines, demonstrating its efficacy.
翻译:基因反转网络(GAN)培训的不稳定性常常归因于梯度下降。 因此,最近的方法旨在调整模型和培训程序以稳定离散更新。 相反,我们研究了GAN培训引起的连续时间动态。理论和玩具实验都表明这些动态事实上是惊人稳定的。从这个角度,我们假设培训GAN的不稳定性产生于连续动态分离的整合错误。我们实验性地核查了众所周知的ODE解答器(如Runge-Kutta)能够稳定培训――当与控制整合错误的正规化器相结合时。我们的方法与通常使用制约功能空间的适应性优化和稳定化技术(如Spectural Algalization)的以往方法大相径庭。我们对CIFAR-10和图像网络的评估表明,我们的方法超越了几个强大的基线,显示了其功效。