Recent studies show strong generative performance in domain translation especially by using transfer learning techniques on the unconditional generator. However, the control between different domain features using a single model is still challenging. Existing methods often require additional models, which is computationally demanding and leads to unsatisfactory visual quality. In addition, they have restricted control steps, which prevents a smooth transition. In this paper, we propose a new approach for high-quality domain translation with better controllability. The key idea is to preserve source features within a disentangled subspace of a target feature space. This allows our method to smoothly control the degree to which it preserves source features while generating images from an entirely new domain using only a single model. Our extensive experiments show that the proposed method can produce more consistent and realistic images than previous works and maintain precise controllability over different levels of transformation. The code is available at https://github.com/LeeDongYeun/FixNoise.
翻译:近期的研究表明,使用条件生成模型的迁移学习技术在领域转换方面表现出强的生成能力,但是单个模型内不同领域特征之间的控制仍然非常具有挑战性。现有方法通常需要额外的模型,这对计算资源需求很高,而且会导致图像的质量不尽如人意。此外,它们控制的步骤通常比较固定,无法实现平滑转换。本文提出了一种实现高质量领域转换且具有更好可控性的新方法。其核心思想是将源特征保留在目标特征空间中的分离子空间中。这使得我们的方法在只使用单个模型的情况下,从全新领域生成图像时,轻松控制其保留源特征的程度。我们的实验结果表明,与之前的工作相比,所提出的方法可以产生更一致和逼真的图像,并且在不同转换级别上保持精确的可控性。代码可在https://github.com/LeeDongYeun/FixNoise中获得。