Recently, attention mechanisms have been explored with ConvNets, both across the spatial and channel dimensions. However, from our knowledge, all the existing methods devote the attention modules to capture local interactions from a uni-scale. In this paper, we propose a Previous Knowledge Channel Attention Module(PKCAM), that captures channel-wise relations across different layers to model the global context. Our proposed module PKCAM is easily integrated into any feed-forward CNN architectures and trained in an end-to-end fashion with a negligible footprint due to its lightweight property. We validate our novel architecture through extensive experiments on image classification and object detection tasks with different backbones. Our experiments show consistent improvements in performances against their counterparts. Our code is published at https://github.com/eslambakr/EMCA.
翻译:最近,在空间和频道两个层面都与ConvNet探索了关注机制,然而,据我们所知,所有现有方法都把关注模块用于从单尺度中捕捉本地互动。在本文件中,我们提议了先前的知识频道关注模块(PKCAM),该模块收集不同层次的渠道关系,以模拟全球背景。我们提议的PKCAM模块很容易纳入任何有线电视新闻网的反馈式结构,并且由于它的轻重属性,其端至端的足迹微乎其微。我们通过对图像分类和不同骨干物体探测任务的广泛实验来验证我们的新结构。我们的实验显示,与对应方相比,我们的工作表现不断改善。我们的代码在https://github.com/eslambakr/EMCA上公布。