The key challenge in learning dense correspondences lies in the lack of ground-truth matches for real image pairs. While photometric consistency losses provide unsupervised alternatives, they struggle with large appearance changes, which are ubiquitous in geometric and semantic matching tasks. Moreover, methods relying on synthetic training pairs often suffer from poor generalisation to real data. We propose Warp Consistency, an unsupervised learning objective for dense correspondence regression. Our objective is effective even in settings with large appearance and view-point changes. Given a pair of real images, we first construct an image triplet by applying a randomly sampled warp to one of the original images. We derive and analyze all flow-consistency constraints arising between the triplet. From our observations and empirical results, we design a general unsupervised objective employing two of the derived constraints. We validate our warp consistency loss by training three recent dense correspondence networks for the geometric and semantic matching tasks. Our approach sets a new state-of-the-art on several challenging benchmarks, including MegaDepth, RobotCar and TSS. Code and models will be released at https://github.com/PruneTruong/DenseMatching.
翻译:学习密度高的函文的关键挑战在于缺乏真实图像配对的地面-真相匹配。 光度一致性损失提供了不受监督的替代物。 虽然光度一致性损失提供了不受监督的替代物, 但它们在巨大的外观变化中挣扎, 这些外观变化在几何和语义匹配任务中是无处不在的。 此外, 依赖合成培训配对的方法往往缺乏对真实数据的概括性。 我们提出Warp Consisticent, 这是一种不受监督的学习目标, 用于密集的函文回归。 我们的目标是在外观和视图点变化巨大的情况下也有效。 如果有一副真实图像, 我们首先通过对原始图像中的一幅随机抽样转折, 来构建一个图像三重体。 我们从观察和实证结果中, 设计出一个通用的、 不受监督的目标, 使用两种衍生的制约。 我们通过培训最近三个密集的几何和语义匹配任务通信网络, 来验证我们的战争一致性损失。 我们的方法在几个挑战的基准上设置了新的状态, 包括Megadepticar 和 TSSrung 和模型。 。 将发布 http://truggsmus/ 。