Many animals and humans process the visual field with a varying spatial resolution (foveated vision) and use peripheral processing to make eye movements and point the fovea to acquire high-resolution information about objects of interest. This architecture results in computationally efficient rapid scene exploration. Recent progress in vision Transformers has brought about new alternatives to the traditionally convolution-reliant computer vision systems. However, these models do not explicitly model the foveated properties of the visual system nor the interaction between eye movements and the classification task. We propose foveated Transformer (FoveaTer) model, which uses pooling regions and saccadic movements to perform object classification tasks using a vision Transformer architecture. Our proposed model pools the image features using squared pooling regions, an approximation to the biologically-inspired foveated architecture, and uses the pooled features as an input to a Transformer Network. It decides on the following fixation location based on the attention assigned by the Transformer to various locations from previous and present fixations. The model uses a confidence threshold to stop scene exploration, allowing to dynamically allocate more fixation/computational resources to more challenging images. We construct an ensemble model using our proposed model and unfoveated model, achieving an accuracy 1.36% below the unfoveated model with 22% computational savings. Finally, we demonstrate our model's robustness against adversarial attacks, where it outperforms the unfoveated model.
翻译:许多动物和人类以不同的空间分辨率(节能视觉)处理视觉场,并使用边缘处理来进行视觉运动,并用边缘处理来显示视觉运动,以显示有关对象的高分辨率信息。这一结构导致计算高效的快速场景勘探。视觉变异器最近的进展为传统上以革命为依存的计算机视觉系统带来了新的替代物。然而,这些模型没有明确模拟视觉系统的顶部特性,或眼运动与分类任务之间的互动。我们提议了变异变异变异变异器(FoveaTer)模型,该模型利用视觉变异器结构来进行视觉运动和分解,以获得高分辨率的物体分类任务。我们提议的模型将图像特征集中到平方集合区域,接近生物激发的织变异结构,并将集合的特性用作对传统变异器网络的投入。它根据变异器对以前和现在的模型不同地点的注意,决定了以下固定位置。模型使用信任阈值来停止现场探索,从而能够以动态的方式将更多的非固定/剖变模型资源分配到更具挑战性的图像中。我们提出的模型,最后用一个模型来显示我们下面的模型。