Generative Adversarial Networks (GANs) can generate near photo realistic images in narrow domains such as human faces. Yet, modeling complex distributions of datasets such as ImageNet and COCO-Stuff remains challenging in unconditional settings. In this paper, we take inspiration from kernel density estimation techniques and introduce a non-parametric approach to modeling distributions of complex datasets. We partition the data manifold into a mixture of overlapping neighborhoods described by a datapoint and its nearest neighbors, and introduce a model, called instance-conditioned GAN (IC-GAN), which learns the distribution around each datapoint. Experimental results on ImageNet and COCO-Stuff show that IC-GAN significantly improves over unconditional models and unsupervised data partitioning baselines. Moreover, we show that IC-GAN can effortlessly transfer to datasets not seen during training by simply changing the conditioning instances, and still generate realistic images. Finally, we extend IC-GAN to the class-conditional case and show semantically controllable generation and competitive quantitative results on ImageNet; while improving over BigGAN on ImageNet-LT. Code and trained models to reproduce the reported results are available at https://github.com/facebookresearch/ic_gan.
翻译:模拟图像网和COCO-Stuff等数据集的复杂分布在无条件的环境下仍然具有挑战性。在本文中,我们从内核密度估计技术中得到灵感,并对复杂数据集的分布采用非参数性的方法。我们将数据元数分成一个由数据点及其近邻描述的重叠邻里混合体,并引入一个模型,称为例定型GAN(IC-GAN),该模型学习每个数据点周围的分布。图像网和COCO-Stuff的实验结果显示,IC-GAN大大改进了无条件模型和不受监督的数据分隔基线。此外,我们表明,IC-GAN可以不费力地将数据转换到在培训过程中看不到的数据集,只需改变调节场,仍然产生现实的图像。最后,我们将IC-GAN(IC-GAN)推广到等级-修饰性案例,并在图像网上显示可操作的生成和竞争性定量结果。在改进BigGAN-LT/LARC/Simagenet上,同时对BAGAN-LTLT/LT在图像复制结果上进行了培训。