One of the issues faced in training Generative Adversarial Nets (GANs) and their variants is the problem of mode collapse, wherein the training stability in terms of the generative loss increases as more training data is used. In this paper, we propose an alternative architecture via the Low-Complexity Neural Network (LCNN), which attempts to learn models with low complexity. The motivation is that controlling model complexity leads to models that do not overfit the training data. We incorporate the LCNN loss function for GANs, Deep Convolutional GANs (DCGANs) and Spectral Normalized GANs (SNGANs), in order to develop hybrid architectures called the LCNN-GAN, LCNN-DCGAN and LCNN-SNGAN respectively. On various large benchmark image datasets, we show that the use of our proposed models results in stable training while avoiding the problem of mode collapse, resulting in better training stability. We also show how the learning behavior can be controlled by a hyperparameter in the LCNN functional, which also provides an improved inception score.
翻译:培训 " 生成反转网 " 及其变体所面临的问题之一是模式崩溃问题,即随着更多培训数据的使用,培训在基因损失方面的稳定性会随着培训数据的使用而提高。在本文件中,我们提议通过低复杂度神经网络(LCNN)建立一个替代结构,试图以低复杂度学习模型。其动机是控制模型的复杂性导致模型不过分适合培训数据。我们纳入了GANs、深革命GANs(DCGANs)和光谱普通GANs(SNGANs)的LCNN损失函数,以便分别开发称为LCNN-GAN、LCNN-DCGAN和LCNN-SNAN的混合结构。关于各种大型基准图像数据集,我们表明,使用我们提议的模型可以带来稳定的培训,同时避免模式崩溃问题,导致更好的培训稳定性。我们还表明,学习行为如何由LCNN功能中的超参数控制,这也提供了更好的初始分数。