Attention mechanism has been regarded as an advanced technique to capture long-range feature interactions and to boost the representation capability for convolutional neural networks. However, we found two ignored problems in current attentional activations-based models: the approximation problem and the insufficient capacity problem of the attention maps. To solve the two problems together, we initially propose an attention module for convolutional neural networks by developing an AW-convolution, where the shape of attention maps matches that of the weights rather than the activations. Our proposed attention module is a complementary method to previous attention-based schemes, such as those that apply the attention mechanism to explore the relationship between channel-wise and spatial features. Experiments on several datasets for image classification and object detection tasks show the effectiveness of our proposed attention module. In particular, our proposed attention module achieves 1.00% Top-1 accuracy improvement on ImageNet classification over a ResNet101 baseline and 0.63 COCO-style Average Precision improvement on the COCO object detection on top of a Faster R-CNN baseline with the backbone of ResNet101-FPN. When integrating with the previous attentional activations-based models, our proposed attention module can further increase their Top-1 accuracy on ImageNet classification by up to 0.57% and COCO-style Average Precision on the COCO object detection by up to 0.45. Code and pre-trained models will be publicly available.
翻译:关注机制被视为一种先进技术,可以捕捉长距离地物相互作用,提高进化神经网络的代表性能力。然而,我们发现当前关注启动模式中有两个被忽视的问题:近似问题和关注地图能力不足问题。为了共同解决这两个问题,我们最初建议为进化神经网络提供一个关注模块,开发AW-演化,关注图的形状与重力而非激活相匹配。我们提议的关注模块是对先前关注机制的一种补充方法,例如那些采用关注机制探索频道与空间特征之间关系的方案。关于图像分类和对象探测任务的若干数据集的实验显示了我们拟议的关注模块的有效性。特别是,我们提议的关注模块在ResNet101基线上实现了图像网络分类1.0%的顶部1级精确度,而在COCOCO的物体探测模型上,在快速R-CNN基线和ResNet101-FPN的主干线上,在与先前的焦点启动目标定位模型相结合时,CO-1010-FN之间的平均分辨率模型将进一步增加我们提议的图像网络的顶部1级精确度。