Convolutional neural networks have allowed remarkable advances in single image super-resolution (SISR) over the last decade. Among recent advances in SISR, attention mechanisms are crucial for high performance SR models. However, few works really discuss why attention works and how it works. In this work, we attempt to quantify and visualize the static attention mechanisms and show that not all attention modules are equally beneficial. We then propose attention in attention network (A$^2$N) for highly accurate image SR. Specifically, our A$^2$N consists of a non-attention branch and a coupling attention branch. Attention dropout module is proposed to generate dynamic attention weights for these two branches based on input features that can suppress unwanted attention adjustments. This allows attention modules to specialize to beneficial examples without otherwise penalties and thus greatly improve the capacity of the attention network with little parameter overhead. Experiments have demonstrated that our model could achieve superior trade-off performances comparing with state-of-the-art lightweight networks. Experiments on local attribution maps also prove attention in attention (A$^2$) structure can extract features from a wider range.
翻译:过去十年来,在单一图像超分辨率(SISR)方面取得了显著的进步,在SISSR最近的进步中,关注机制对高性能SR模型至关重要,然而,几乎没有什么工作真正讨论为什么关注起作用和如何发挥作用。在这项工作中,我们试图量化和想象静态关注机制,并表明并非所有关注模块都具有同等好处。然后我们建议关注网络(A$2N)对非常准确的图像SR给予关注。具体地说,我们的A$2N由一个不注意分支和一个混合关注分支组成。建议关注退出模块根据可抑制不必要关注调整的投入特征为这两个分支带来动态关注权重。这使得关注模块能够专门关注有益的实例,而无需以其他方式处罚,从而大大提高关注网络的能力,同时使用微小的参数中位。实验表明,我们的模型能够实现与州级轻量网络相比较的更高交易性表现。对本地归属图的实验也证明了关注度(A$2$)结构中的特征。