Current image-to-image translation methods formulate the task with conditional generation models, leading to learning only the recolorization or regional changes as being constrained by the rich structural information provided by the conditional contexts. In this work, we propose introducing the vector quantization technique into the image-to-image translation framework. The vector quantized content representation can facilitate not only the translation, but also the unconditional distribution shared among different domains. Meanwhile, along with the disentangled style representation, the proposed method further enables the capability of image extension with flexibility in both intra- and inter-domains. Qualitative and quantitative experiments demonstrate that our framework achieves comparable performance to the state-of-the-art image-to-image translation and image extension methods. Compared to methods for individual tasks, the proposed method, as a unified framework, unleashes applications combining image-to-image translation, unconditional generation, and image extension altogether. For example, it provides style variability for image generation and extension, and equips image-to-image translation with further extension capabilities.
翻译:当前图像到图像翻译方法以有条件生成模型来制定任务,导致只学习受有条件背景提供的丰富结构信息制约的变色化或区域变化。 在这项工作中,我们提议将矢量量化技术引入图像到图像翻译框架。矢量量化内容表达方式不仅可以促进翻译,还可以促进在不同领域之间无条件共享。同时,与分解的风格表达方式一道,拟议方法进一步使得图像扩展能力在内部和内部都具有灵活性。定性和定量实验表明,我们的框架实现了与最新图像到图像翻译和图像扩展方法相似的性能。与单个任务的方法相比,拟议方法作为一个统一框架,可以启动将图像到图像翻译、无条件生成和图像扩展相结合的应用。例如,它为图像生成和扩展提供样式变异性,并为图像到图像转换提供进一步的扩展能力。