We study on image super-resolution (SR), which aims to recover realistic textures from a low-resolution (LR) image. Recent progress has been made by taking high-resolution images as references (Ref), so that relevant textures can be transferred to LR images. However, existing SR approaches neglect to use attention mechanisms to transfer high-resolution (HR) textures from Ref images, which limits these approaches in challenging cases. In this paper, we propose a novel Texture Transformer Network for Image Super-Resolution (TTSR), in which the LR and Ref images are formulated as queries and keys in a transformer, respectively. TTSR consists of four closely-related modules optimized for image generation tasks, including a learnable texture extractor by DNN, a relevance embedding module, a hard-attention module for texture transfer, and a soft-attention module for texture synthesis. Such a design encourages joint feature learning across LR and Ref images, in which deep feature correspondences can be discovered by attention, and thus accurate texture features can be transferred. The proposed texture transformer can be further stacked in a cross-scale way, which enables texture recovery from different levels (e.g., from 1x to 4x magnification). Extensive experiments show that TTSR achieves significant improvements over state-of-the-art approaches on both quantitative and qualitative evaluations.
翻译:我们研究的是图像超分辨率(SR),目的是从低分辨率(LR)图像中恢复现实的纹理。最近的进展是通过将高分辨率图像作为参考(Ref)而取得的,这样可以将相关纹理转换到LR图像。然而,现有的SR方法忽视了使用关注机制将Ref图像的高分辨率(HR)纹理传输到高分辨率(HR)图理,这限制了这些在具有挑战性的情况下采用的方法。在本文件中,我们建议建立一个新型图像超分辨率(TTTSR)图象转换网络(TTSR),其中LR和Ref图像分别作为变异器中的查询和密钥制作。TSSR由四个为图像生成任务优化的密切相关的模块组成,其中包括DNNE的可学习质素提取模块、相关嵌入模块、文本传输的硬性模块,以及文本合成软性模块。这种设计鼓励在LR和Ref图像中联合学习特征,通过关注来发现深度特征通信,从而可以将精确的文字特征转换成。拟议的文本变异式变式模型可以进一步从4级到升级。