Camera calibration is integral to robotics and computer vision algorithms that seek to infer geometric properties of the scene from visual input streams. In practice, calibration is a laborious procedure requiring specialized data collection and careful tuning. This process must be repeated whenever the parameters of the camera change, which can be a frequent occurrence for mobile robots and autonomous vehicles. In contrast, self-supervised depth and ego-motion estimation approaches can bypass explicit calibration by inferring per-frame projection models that optimize a view synthesis objective. In this paper, we extend this approach to explicitly calibrate a wide range of cameras from raw videos in the wild. We propose a learning algorithm to regress per-sequence calibration parameters using an efficient family of general camera models. Our procedure achieves self-calibration results with sub-pixel reprojection error, outperforming other learning-based methods. We validate our approach on a wide variety of camera geometries, including perspective, fisheye, and catadioptric. Finally, we show that our approach leads to improvements in the downstream task of depth estimation, achieving state-of-the-art results on the EuRoC dataset with greater computational efficiency than contemporary methods.
翻译:相机校准是机器人和计算机视觉算法不可分割的一部分,这些算法试图从视觉输入流中推断场景的几何特性。在实践中,校准是一种艰巨的程序,需要专门的数据收集和仔细的调整。每当相机变化的参数发生时,这一过程必须重复,这可能是移动机器人和自主飞行器的常见现象。相比之下,自我监督的深度和自动估计方法可以绕过明确的校准,方法是通过推断每框架的预测模型,优化观景合成目标。在本文中,我们扩展了这一方法,以明确校准野生生生视频中的各种相机。我们建议采用学习的算法,利用一个高效的普通相机模型组合来反向每个序列校准参数。我们的程序通过子像素再预测错误实现自我校准结果,优于其他基于学习的方法。我们验证了我们对于各种摄影机地理模型的方法,包括视角、鱼眼和卡塔迪氏仪。最后,我们的方法显示,我们的方法可以改进下游的深度估计任务,实现实时数据测算法的效率,从而提高EuC的进度。