Style transfer is the task of reproducing the semantic contents of a source image in the artistic style of a second target image. In this paper, we present NeAT, a new state-of-the art feed-forward style transfer method. We re-formulate feed-forward style transfer as image editing, rather than image generation, resulting in a model which improves over the state-of-the-art in both preserving the source content and matching the target style. An important component of our model's success is identifying and fixing "style halos", a commonly occurring artefact across many style transfer techniques. In addition to training and testing on standard datasets, we introduce the BBST-4M dataset, a new, large scale, high resolution dataset of 4M images. As a component of curating this data, we present a novel model able to classify if an image is stylistic. We use BBST-4M to improve and measure the generalization of NeAT across a huge variety of styles. Not only does NeAT offer state-of-the-art quality and generalization, it is designed and trained for fast inference at high resolution.
翻译:风格转移是将源图像的语义内容复制到第二个目标图像的艺术风格中的任务。在本文中,我们介绍了NeAT,一种新的最先进的前馈风格转移方法。我们将前馈风格转移重新定义为图像编辑而不是图像生成,从而得到一种模型,它在保留源内容和匹配目标风格方面都优于现有技术的水平。我们模型成功的一个重要组成部分是发现和修复“风格光环”,这是许多风格转移技术共同出现的一种现象。除了在标准数据集上进行训练和测试外,我们还介绍了BBST-4M数据集,这是一个新的、大规模、高分辨率的4M图像数据集。作为策划这个数据的一部分,我们介绍了一种新颖的模型,能够对图像进行风格分类。我们使用BBST-4M来改进和衡量NeAT在各种风格上的泛化能力。NeAT不仅提供了最先进的质量和泛化能力,而且旨在在高分辨率下进行快速推理。