In humans, Attention is a core property of all perceptual and cognitive operations. Given our limited ability to process competing sources, attention mechanisms select, modulate, and focus on the information most relevant to behavior. For decades, concepts and functions of attention have been studied in philosophy, psychology, neuroscience, and computing. For the last six years, this property has been widely explored in deep neural networks. Currently, the state-of-the-art in Deep Learning is represented by neural attention models in several application domains. This survey provides a comprehensive overview and analysis of developments in neural attention models. We systematically reviewed hundreds of architectures in the area, identifying and discussing those in which attention has shown a significant impact. We also developed and made public an automated methodology to facilitate the development of reviews in the area. By critically analyzing 650 works, we describe the primary uses of attention in convolutional, recurrent networks and generative models, identifying common subgroups of uses and applications. Furthermore, we describe the impact of attention in different application domains and their impact on neural networks' interpretability. Finally, we list possible trends and opportunities for further research, hoping that this review will provide a succinct overview of the main attentional models in the area and guide researchers in developing future approaches that will drive further improvements.
翻译:在人类中,注意力是所有感知和认知活动的核心属性。鉴于我们处理相互竞争的来源的能力有限,关注机制选择、调整和侧重于与行为最相关的信息。几十年来,在哲学、心理学、神经科学和计算机方面研究了关注的概念和功能。在过去六年里,这种财产在深层神经网络中得到了广泛的探索。目前,深层学习中最先进的是若干应用领域的神经关注模型。这项调查对神经关注模型的发展动态进行了全面的概述和分析。我们系统地审查了该地区数百个结构,查明和讨论关注已产生重大影响的那些结构。我们还开发并公布了一种自动化方法,以促进该领域审查的发展。我们通过批判性地分析650项作品,描述了革命性、经常性网络和基因化模型的主要关注用途,确定了共同的使用和应用分组。此外,我们描述了不同应用领域关注的影响及其对神经网络解释能力的影响。最后,我们列出了进一步研究的可能趋势和机会,希望这一审查将指导未来研究领域的主要关注方向。