Recent style transfer problems are still largely dominated by Generative Adversarial Network (GAN) from the perspective of cross-domain image-to-image (I2I) translation, where the pivotal issue is to learn and transfer target-domain style patterns onto source-domain content images. This paper handles the problem of translating real pictures into traditional Chinese ink-wash paintings, i.e., Chinese ink-wash painting style transfer. Though a wide range of I2I models tackle this problem, a notable challenge is that the content details of the source image could be easily erased or corrupted due to the transfer of ink-wash style elements. To remedy this issue, we propose to incorporate saliency detection into the unpaired I2I framework to regularize image content, where the detected saliency map is utilized from two aspects: (\romannumeral1) we propose saliency IOU (SIOU) loss to explicitly regularize object content structure by enforcing saliency consistency before and after image stylization; (\romannumeral2) we propose saliency adaptive normalization (SANorm) which implicitly enhances object structure integrity of the generated paintings by dynamically injecting image saliency information into the generator to guide stylization process. Besides, we also propose saliency attended discriminator which harnesses image saliency information to focus generative adversarial attention onto the drawn objects, contributing to generating more vivid and delicate brush strokes and ink-wash textures. Extensive qualitative and quantitative experiments demonstrate superiority of our approach over related advanced image stylization methods in both GAN and diffusion model paradigms.
翻译:当前风格迁移问题仍主要从跨域图像到图像(I2I)转换的角度由生成对抗网络(GAN)主导,其核心在于学习目标域风格模式并将其迁移至源域内容图像。本文处理将真实照片转换为传统中国水墨画的问题,即中国水墨画风格迁移。尽管已有多种I2I模型应对此问题,但一个显著挑战是源图像的内容细节可能因水墨风格元素的迁移而轻易被抹除或破坏。为解决此问题,我们提出在非配对I2I框架中引入显著性检测以正则化图像内容,其中检测到的显著性图从两方面被利用:(一)我们提出显著性交并比(SIOU)损失,通过强制图像风格化前后的显著性一致性来显式正则化对象内容结构;(二)我们提出显著性自适应归一化(SANorm),通过动态向生成器注入图像显著性信息以指导风格化过程,从而隐式增强生成画作的对象结构完整性。此外,我们还提出显著性注意力判别器,利用图像显著性信息将生成对抗注意力聚焦于绘制对象,有助于生成更生动细腻的笔触与水墨纹理。广泛的定性与定量实验表明,在GAN和扩散模型两种范式下,我们的方法均优于相关先进的图像风格化方法。