Homography estimation is often an indispensable step in many computer vision tasks. The existing approaches, however, are not robust to illumination and/or larger viewpoint changes. In this paper, we propose bidirectional implicit Homography Estimation (biHomE) loss for unsupervised homography estimation. biHomE minimizes the distance in the feature space between the warped image from the source viewpoint and the corresponding image from the target viewpoint. Since we use a fixed pre-trained feature extractor and the only learnable component of our framework is the homography network, we effectively decouple the homography estimation from representation learning. We use an additional photometric distortion step in the synthetic COCO dataset generation to better represent the illumination variation of the real-world scenarios. We show that biHomE achieves state-of-the-art performance on synthetic COCO dataset, which is also comparable or better compared to supervised approaches. Furthermore, the empirical results demonstrate the robustness of our approach to illumination variation compared to existing methods.
翻译:许多计算机的视觉任务通常都必须采取对同系物的估计,但现有的方法对照明和(或)更大的观点变化并不有力。在本文中,我们提议对未经监督的同系物估计进行双向隐含同系物估计(BiHomE)损失。双HomE最大限度地缩小了从源角度扭曲的图像与从目标角度对相应图像之间在特征空间上的距离。由于我们使用固定的预先训练的特征提取器,而且我们框架的唯一可学习部分是同系物网络,我们有效地将同系物估计与代言学习脱钩。我们在合成COCO数据集生成中使用了额外的光度偏差步骤,以更好地代表真实世界情景的无光度变化。我们显示,双HomE在合成COCO数据集上达到最新水平的性能,这与所监督的方法相比也是可比或更好的。此外,经验结果显示,我们处理同现有方法相比,对同系物差异的方法非常稳健。