The past decade has witnessed a drastic increase in modern deep neural networks (DNNs) size, especially for generative adversarial networks (GANs). Since GANs usually suffer from high computational complexity, researchers have shown an increased interest in applying pruning methods to reduce the training and inference costs of GANs. Among different pruning methods invented for supervised learning, dynamic sparse training (DST) has gained increasing attention recently as it enjoys excellent training efficiency with comparable performance to post-hoc pruning. Hence, applying DST on GANs, where we train a sparse GAN with a fixed parameter count throughout training, seems to be a good candidate for reducing GAN training costs. However, a few challenges, including the degrading training instability, emerge due to the adversarial nature of GANs. Hence, we introduce a quantity called balance ratio (BR) to quantify the balance of the generator and the discriminator. We conduct a series of experiments to show the importance of BR in understanding sparse GAN training. Building upon single dynamic sparse training (SDST), where only the generator is adjusted during training, we propose double dynamic sparse training (DDST) to control the BR during GAN training. Empirically, DDST automatically determines the density of the discriminator and greatly boosts the performance of sparse GANs on multiple datasets.
翻译:过去十年来,现代深层神经网络(DNNS)的规模急剧增加,特别是在基因对抗网络(GANs)中,现代深层神经网络(DNNS)的规模急剧增加。由于GANs通常具有很高的计算复杂度,因此研究人员对采用修剪方法减少GANs的培训和推算费用表现出越来越大的兴趣。在为监督学习而发明的不同修剪方法中,动态稀少的培训最近日益受到越来越多的注意,因为它具有与后热处理的类似性能,因此,在GANs上应用DST(DST),我们在那里培训一个有固定参数计数的稀少的GAN,这似乎是降低GAN培训费用的好人选。然而,由于GANs的对抗性质,包括降低的培训不稳定性,我们提出了数量上所谓的平衡率(BR),以量化发电机和导师的平衡性能。我们进行了一系列实验,以表明BR在理解稀少GAN培训中的重要性。在单项动态稀疏培训(SDST)的基础上,只有发电机在培训过程中进行调整,我们提议将GAN的高度动态稀薄性培训自动地推进GNDDDDDD(D)到GNDDDDDDD),因此,我们提议在GAN的多重性训练期间将GDDDDDDDDDDDDDDDDD(D)自动地进行双重的高度的升级的密度的升级。</s>