This work aims at transferring a Generative Adversarial Network (GAN) pre-trained on one image domain to a new domain referring to as few as just one target image. The main challenge is that, under limited supervision, it is extremely difficult to synthesize photo-realistic and highly diverse images, while acquiring representative characters of the target. Different from existing approaches that adopt the vanilla fine-tuning strategy, we import two lightweight modules to the generator and the discriminator respectively. Concretely, we introduce an attribute adaptor into the generator yet freeze its original parameters, through which it can reuse the prior knowledge to the most extent and hence maintain the synthesis quality and diversity. We then equip the well-learned discriminator backbone with an attribute classifier to ensure that the generator captures the appropriate characters from the reference. Furthermore, considering the poor diversity of the training data (i.e., as few as only one image), we propose to also constrain the diversity of the generative domain in the training process, alleviating the optimization difficulty. Our approach brings appealing results under various settings, substantially surpassing state-of-the-art alternatives, especially in terms of synthesis diversity. Noticeably, our method works well even with large domain gaps, and robustly converges within a few minutes for each experiment.
翻译:这项工作旨在将在一个图像域上经过预先训练的基因反反影网络(GAN)转移到一个新域,仅仅仅一个目标图像,主要挑战在于,在有限的监督下,很难合成照片现实和高度多样化的图像,同时获得目标的代表性字符。不同于采用香草微调战略的现有方法,我们向发电机和导体分别输入两个轻量模块。具体地说,我们在发电机中引入一个属性适配器,同时冻结其原始参数,通过这些参数,它可以将先前的知识再利用到最大程度,从而保持合成质量和多样性。然后,我们用一个属性分类仪来为学习良好的歧视者骨干配备一个属性分类仪,以确保生成者从参考中捕捉到适当的字符。此外,考虑到培训数据的多样性差(即,只有很少一个图像),我们建议同时限制培训过程中的基因化领域的多样性,减轻优化难度。我们的方法在各种环境中带来吸引的结果,甚至大大超过最先进的替代方法,特别是在合成多样性方面。值得注意的是,我们的方法在每一个领域内都进行紧密的整合。