Image-to-image (I2I) translation is usually carried out among discrete domains. However, image domains, often corresponding to a physical value, are usually continuous. In other words, images gradually change with the value, and there exists no obvious gap between different domains. This paper intends to build the model for I2I translation among continuous varying domains. We first divide the whole domain coverage into discrete intervals, and explicitly model the latent style code for the center of each interval. To deal with continuous translation, we design the editing modules, changing the latent style code along two directions. These editing modules help to constrain the codes for domain centers during training, so that the model can better understand the relation among them. To have diverse results, the latent style code is further diversified with either the random noise or features from the reference image, giving the individual style code to the decoder for label-based or reference-based synthesis. Extensive experiments on age and viewing angle translation show that the proposed method can achieve high-quality results, and it is also flexible for users.
翻译:图像到图像翻译( I2I) 通常在离散域间进行。 但是, 图像域, 通常与物理值相对应, 通常是连续的。 换句话说, 图像随着值逐渐变化, 不同域间没有明显差距 。 本文打算将I2I 翻译模型建在连续的不同域间间。 我们首先将整个域覆盖分成离散间隔, 并明确为每个间隔中心的潜在样式代码建模。 处理连续翻译, 我们设计编辑模块, 沿着两个方向修改潜在样式代码。 这些编辑模块有助于限制域中心的代码, 以便让模型更好地了解它们之间的关系。 要取得不同的结果, 潜在样式代码会随着引用图像的随机噪音或特征而进一步多样化, 将单个样式代码赋予基于标签或基于参考的合成的解码。 年龄和角度翻译的广泛实验显示, 拟议的方法可以取得高质量结果, 用户也会灵活。