A variety of attention mechanisms have been studied to improve the performance of various computer vision tasks. However, the prior methods overlooked the significance of retaining the information on both channel and spatial aspects to enhance the cross-dimension interactions. Therefore, we propose a global attention mechanism that boosts the performance of deep neural networks by reducing information reduction and magnifying the global interactive representations. We introduce 3D-permutation with multilayer-perceptron for channel attention alongside a convolutional spatial attention submodule. The evaluation of the proposed mechanism for the image classification task on CIFAR-100 and ImageNet-1K indicates that our method stably outperforms several recent attention mechanisms with both ResNet and lightweight MobileNet.
翻译:研究各种关注机制以改善各种计算机愿景任务的业绩,但先前的方法忽略了保留频道和空间信息以加强跨层互动的重要性,因此,我们建议建立一个全球关注机制,通过减少信息减少和扩大全球互动演示,提高深神经网络的性能。我们引入了三维调和多层感应器,供频道关注,并配有动态空间关注子模块。对拟议的CIFAR-100和图像网络-1K图像分类任务机制的评估表明,我们的方法在ResNet和轻量级移动网络上都明显优于最近几个关注机制。