Recently, deep learning has been successfully applied to the single-image super-resolution (SISR) with remarkable performance. However, most existing methods focus on building a more complex network with a large number of layers, which can entail heavy computational costs and memory storage. To address this problem, we present a lightweight Self-Calibrated Efficient Transformer (SCET) network to solve this problem. The architecture of SCET mainly consists of the self-calibrated module and efficient transformer block, where the self-calibrated module adopts the pixel attention mechanism to extract image features effectively. To further exploit the contextual information from features, we employ an efficient transformer to help the network obtain similar features over long distances and thus recover sufficient texture details. We provide comprehensive results on different settings of the overall network. Our proposed method achieves more remarkable performance than baseline methods. The source code and pre-trained models are available at https://github.com/AlexZou14/SCET.
翻译:最近,对单一图像超分辨率(SISR)成功地应用了深层次学习,其性能显著,然而,大多数现有方法侧重于建立一个更复杂的网络,其层层众多,可能带来沉重的计算成本和内存存储。为解决这一问题,我们提出了一个轻量的自我校准高效变异器(SCET)网络来解决这一问题。标准电子技术的设计主要由自我校准模块和高效变压器块组成,在这种结构中,自我校准的模块采用像素关注机制来有效提取图像特征。为了进一步利用各种特征的相貌信息,我们使用高效变异器来帮助网络获得长距离的类似特征,从而恢复足够的纹理细节。我们在整个网络的不同环境中提供了全面的结果。我们提出的方法比基线方法更能取得显著的性能。源代码和预先培训的模型见https://github.com/AlexZou14/SCET。