Local feature matching aims at finding correspondences between a pair of images. Although current detector-free methods leverage Transformer architecture to obtain an impressive performance, few works consider maintaining local consistency. Meanwhile, most methods struggle with large scale variations. To deal with the above issues, we propose Adaptive Spot-Guided Transformer (ASTR) for local feature matching, which jointly models the local consistency and scale variations in a unified coarse-to-fine architecture. The proposed ASTR enjoys several merits. First, we design a spot-guided aggregation module to avoid interfering with irrelevant areas during feature aggregation. Second, we design an adaptive scaling module to adjust the size of grids according to the calculated depth information at fine stage. Extensive experimental results on five standard benchmarks demonstrate that our ASTR performs favorably against state-of-the-art methods. Our code will be released on https://astr2023.github.io.
翻译:局部特征匹配的目标是在一对图像之间寻找对应关系。虽然当前的无检测器方法利用Transformer架构获得了惊人的性能,但很少有研究考虑维护局部一致性。与此同时,大多数方法都无法处理大规模变化。为了解决以上问题,我们提出了自适应聚斑Transformer(ASTR)进行局部特征匹配,它在一个统一的从粗到细的架构中联合建模局部一致性和尺度变化。所提出的ASTR具有以下优点。首先,我们设计了一种聚斑聚合模块,以避免在特征聚合过程中干扰无关区域。其次,我们设计了一种自适应缩放模块,根据在精细阶段计算的深度信息调整网格的大小。在五个标准基准测试上进行的大量实验结果表明,我们的ASTR表现优于最先进的方法。我们的代码将在https://astr2023.github.io上发布。