Deep learning based approaches has achieved great performance in single image super-resolution (SISR). However, recent advances in efficient super-resolution focus on reducing the number of parameters and FLOPs, and they aggregate more powerful features by improving feature utilization through complex layer connection strategies. These structures may not be necessary to achieve higher running speed, which makes them difficult to be deployed to resource-constrained devices. In this work, we propose a novel Residual Local Feature Network (RLFN). The main idea is using three convolutional layers for residual local feature learning to simplify feature aggregation, which achieves a good trade-off between model performance and inference time. Moreover, we revisit the popular contrastive loss and observe that the selection of intermediate features of its feature extractor has great influence on the performance. Besides, we propose a novel multi-stage warm-start training strategy. In each stage, the pre-trained weights from previous stages are utilized to improve the model performance. Combined with the improved contrastive loss and training strategy, the proposed RLFN outperforms all the state-of-the-art efficient image SR models in terms of runtime while maintaining both PSNR and SSIM for SR. In addition, we won the first place in the runtime track of the NTIRE 2022 efficient super-resolution challenge. Code will be available at https://github.com/fyan111/RLFN.
翻译:深度学习方法在单一图像超分辨率(SISR)中取得了巨大成绩。然而,最近在高效超级分辨率方法方面取得的进步侧重于减少参数和FLOP数量,通过通过复杂层连接战略改进地貌利用功能利用,汇集了更强大的特征;这些结构对于提高运行速度可能并不必要,因此难以将其部署到资源受限制的设备中。在这项工作中,我们提议建立一个全新的剩余地方特质网络(RLFNF),主要想法是利用三个富集层来学习剩余地方特质以简化集成功能,从而在模型性能和推断时间之间实现良好的平衡。此外,我们重新审视了流行的对比损失,并观察到选择其地貌提取器的中间特征对性能有很大影响。此外,我们提出一个新的多阶段热启动培训战略。在每一个阶段,利用前几个阶段的预培训权重来改进模型性能。与改进的对比性损失和培训战略相结合,拟议的RFNFFM超越了所有在运行时运行时最先进的图像SR模型。在运行时将运行时的PSNFR/RMRM 的轨道上,同时将保留RIS/S-CRMU 。