Image super-resolution (SR) is a fast-moving field with novel architectures attracting the spotlight. However, most SR models were optimized with dated training strategies. In this work, we revisit the popular RCAN model and examine the effect of different training options in SR. Surprisingly (or perhaps as expected), we show that RCAN can outperform or match nearly all the CNN-based SR architectures published after RCAN on standard benchmarks with a proper training strategy and minimal architecture change. Besides, although RCAN is a very large SR architecture with more than four hundred convolutional layers, we draw a notable conclusion that underfitting is still the main problem restricting the model capability instead of overfitting. We observe supportive evidence that increasing training iterations clearly improves the model performance while applying regularization techniques generally degrades the predictions. We denote our simply revised RCAN as RCAN-it and recommend practitioners to use it as baselines for future research. Code is publicly available at https://github.com/zudi-lin/rcan-it.
翻译:图像超分辨率(SR)是一个快速移动的领域,其新结构吸引了人们的注意。然而,大多数SR模型都是以过时的培训战略优化的。在这项工作中,我们重新审视了流行的RCAN模型,并考察了斯洛伐克共和国各种培训选项的效果。令人惊讶(或可能如预期的那样),我们显示RCAN能够超越或匹配在RCAN之后出版的几乎所有基于CNN的SR结构,这些结构以适当的培训战略和最低限度的结构变化作为标准基准。此外,虽然RCAN是一个非常庞大的SR结构,拥有400多个卷积层,但我们得出一个显著的结论,即不足仍然是限制模型能力而不是过度适应的主要问题。我们观察到有支持性证据表明,增加培训迭代显然改善了模型的性能,同时使用正规化技术通常会降低预测。我们简单地将RCAN作为RCAN作为RCAN-it,并建议从业者将其用作未来研究的基线。代码可在https://github.com/zudi-lin/rcan-it上公开查阅。