Image-to-Image (I2I) multi-domain translation models are usually evaluated also using the quality of their semantic interpolation results. However, state-of-the-art models frequently show abrupt changes in the image appearance during interpolation, and usually perform poorly in interpolations across domains. In this paper, we propose a new training protocol based on three specific losses which help a translation network to learn a smooth and disentangled latent style space in which: 1) Both intra- and inter-domain interpolations correspond to gradual changes in the generated images and 2) The content of the source image is better preserved during the translation. Moreover, we propose a novel evaluation metric to properly measure the smoothness of latent style space of I2I translation models. The proposed method can be plugged into existing translation approaches, and our extensive experiments on different datasets show that it can significantly boost the quality of the generated images and the graduality of the interpolations.
翻译:图像到图像( I2I) 多域翻译模型通常也使用其语义内插结果的质量来评估。 但是, 最先进的模型经常显示内插期间图像外观的突变, 并且通常在跨域的内插中表现不佳 。 在本文中, 我们基于三个具体损失提出一个新的培训协议, 帮助翻译网络学习一个平滑和分解的潜在风格空间, 其中:(1) 内部和内部的内插图与生成图像的逐渐变化相对应,(2) 翻译期间来源图像的内容得到更好的保存。 此外, 我们提出了一个新的评估指标, 以适当测量I2I翻译模型潜在样式空间的平滑性。 提议的方法可以插入到现有的翻译方法中, 我们在不同的数据集上进行的广泛实验表明, 它可以大大提高生成图像的质量和内插图的渐进性。