Producing diverse and realistic images with generative models such as GANs typically requires large scale training with vast amount of images. GANs trained with extremely limited data can easily overfit to few training samples and display undesirable properties like "stairlike" latent space where transitions in latent space suffer from discontinuity, occasionally yielding abrupt changes in outputs. In this work, we consider the situation where neither large scale dataset of our interest nor transferable source dataset is available, and seek to train existing generative models with minimal overfitting and mode collapse. We propose latent mixup-based distance regularization on the feature space of both a generator and the counterpart discriminator that encourages the two players to reason not only about the scarce observed data points but the relative distances in the feature space they reside. Qualitative and quantitative evaluation on diverse datasets demonstrates that our method is generally applicable to existing models to enhance both fidelity and diversity under the constraint of limited data. Code will be made public.
翻译:以基因模型(如GANs)产生多样化和现实的图像,通常需要用大量图像进行大规模培训。用极其有限的数据培训的GANs可以很容易地取代少数培训样本,并展示出不可取的特性,如潜层空间的过渡不连续,有时会突然导致产出突变。在这项工作中,我们考虑到既无法获得我们感兴趣的大型数据集,也无法获得可转让源数据集,我们寻求以最小的过度和模式崩溃来培训现有的基因模型。我们提议在发电机和对口歧视者的地貌空间进行潜在的混杂式距离规范,鼓励这两个角色不仅说明所观察到的稀少的数据点,而且说明它们所居住的地貌空间的相对距离。对各种数据集的定性和定量评价表明,我们的方法通常适用于现有模型,以便在有限数据的限制下加强忠诚和多样性。我们的方法将会被公之于众。