Transformer trackers have achieved impressive advancements recently, where the attention mechanism plays an important role. However, the independent correlation computation in the attention mechanism could result in noisy and ambiguous attention weights, which inhibits further performance improvement. To address this issue, we propose an attention in attention (AiA) module, which enhances appropriate correlations and suppresses erroneous ones by seeking consensus among all correlation vectors. Our AiA module can be readily applied to both self-attention blocks and cross-attention blocks to facilitate feature aggregation and information propagation for visual tracking. Moreover, we propose a streamlined Transformer tracking framework, dubbed AiATrack, by introducing efficient feature reuse and target-background embeddings to make full use of temporal references. Experiments show that our tracker achieves state-of-the-art performance on six tracking benchmarks while running at a real-time speed.
翻译:最近,在关注机制发挥重要作用的地方,变革跟踪器取得了令人印象深刻的进步,然而,关注机制的独立相关计算可能导致关注权的吵闹和模糊,从而抑制进一步的绩效改进。为了解决这一问题,我们建议关注(AiA)模块,该模块通过在所有相关矢量之间寻求共识,加强适当的关联,抑制错误的关联。我们的AiA模块可以很容易地适用于自关注区块和交叉关注区块,以便利特征汇总和信息传播,以便进行视觉跟踪。此外,我们建议采用一个称为AiATrack的简化变换器跟踪框架,称为AiATrack,采用高效的特性再利用和目标后地嵌入,以充分利用时间参考。实验显示,我们的跟踪器在实时运行的同时,在六个跟踪基准上取得了最新业绩。