We present a technique for estimating the relative 3D rotation of an RGB image pair in an extreme setting, where the images have little or no overlap. We observe that, even when images do not overlap, there may be rich hidden cues as to their geometric relationship, such as light source directions, vanishing points, and symmetries present in the scene. We propose a network design that can automatically learn such implicit cues by comparing all pairs of points between the two input images. Our method therefore constructs dense feature correlation volumes and processes these to predict relative 3D rotations. Our predictions are formed over a fine-grained discretization of rotations, bypassing difficulties associated with regressing 3D rotations. We demonstrate our approach on a large variety of extreme RGB image pairs, including indoor and outdoor images captured under different lighting conditions and geographic locations. Our evaluation shows that our model can successfully estimate relative rotations among non-overlapping images without compromising performance over overlapping image pairs.
翻译:我们提出了一个在极端环境下估计一个 RGB 图像配对的相对三维旋转的技术,因为图像很少或没有重叠。我们观察到,即使图像没有重叠,在几何关系上也可能隐藏着许多隐藏的线索,例如光源方向、消失点和现场的对称。我们提出了一个网络设计,通过比较两个输入图像之间的所有两对点,可以自动学习这种隐含的提示。因此,我们的方法构建了密集的特性相关量,并流程来预测相对的三维旋转。我们的预测是建立在精细的旋转离散之上,绕过与三维回移相关的困难。我们展示了我们对于大量极端 RGB 图像配对的方法,包括在不同的照明条件和地理位置下采集的室内和户外图像。我们的评估表明,我们的模型可以成功地估计非重叠图像之间的相对旋转,而不影响重叠图像配对的性能。