Establishing dense correspondence between two images is a fundamental computer vision problem, which is typically tackled by matching local feature descriptors. However, without global awareness, such local features are often insufficient for disambiguating similar regions. And computing the pairwise feature correlation across images is both computation-expensive and memory-intensive. To make the local features aware of the global context and improve their matching accuracy, we introduce DenseGAP, a new solution for efficient Dense correspondence learning with a Graph-structured neural network conditioned on Anchor Points. Specifically, we first propose a graph structure that utilizes anchor points to provide sparse but reliable prior on inter- and intra-image context and propagates them to all image points via directed edges. We also design a graph-structured network to broadcast multi-level contexts via light-weighted message-passing layers and generate high-resolution feature maps at low memory cost. Finally, based on the predicted feature maps, we introduce a coarse-to-fine framework for accurate correspondence prediction using cycle consistency. Our feature descriptors capture both local and global information, thus enabling a continuous feature field for querying arbitrary points at high resolution. Through comprehensive ablative experiments and evaluations on large-scale indoor and outdoor datasets, we demonstrate that our method advances the state-of-the-art of correspondence learning on most benchmarks.
翻译:在两种图像之间建立密集的对应关系是一个基本的计算机视觉问题,通常通过匹配本地特征描述符来解决。然而,如果没有全球意识,这些本地特征往往不足以掩盖类似的区域。计算图像之间的对称特征相关性既是计算成本的,也是记忆密集型的。为了使本地特征了解全球背景,并提高其匹配准确性,我们引入了DenseGAP,这是与以安氏点为条件的图形结构神经网络进行高效的常识通信学习的新解决方案。具体地说,我们首先提出一个图形结构,利用定位点在图像间和内部背景之前提供稀少但可靠的信息,并通过定向边缘将其传播到所有图像点。我们还设计了一个图形结构网络,通过轻量度信息传递层和记忆层广播多层次环境,并以低记忆成本生成高分辨率特征图。最后,我们根据预测特征图,引入了一个以循环一致性为条件的精确通信网络的粗度到线框架。我们的特点描述着本地和全球信息,从而能够通过定向的深度实验和高分辨率的深度分析,从而能够持续地显示我们所处的深度数据测试和高分辨率的深度分析。