Camera localization is a fundamental and crucial problem for many robotic applications. In recent years, using deep-learning for camera-based localization has become a popular research direction. However, they lack robustness to large domain shifts, which can be caused by seasonal or illumination changes between training and testing data sets. Data augmentation is an attractive approach to tackle this problem, as it does not require additional data to be provided. However, existing augmentation methods blindly perturb all pixels and therefore cannot achieve satisfactory performance. To overcome this issue, we proposed RADA, a system whose aim is to concentrate on perturbing the geometrically informative parts of the image. As a result, it learns to generate minimal image perturbations that are still capable of perplexing the network. We show that when these examples are utilized as augmentation, it greatly improves robustness. We show that our method outperforms previous augmentation techniques and achieves up to two times higher accuracy than the SOTA localization models (e.g., AtLoc and MapNet) when tested on `unseen' challenging weather conditions.
翻译:对于许多机器人应用程序来说,相机本地化是一个根本性和关键的问题。近年来,使用深层学习进行相机本地化已成为一个受欢迎的研究方向。然而,它们缺乏对大域变迁的稳健性,而大域变迁可能是由于培训和测试数据集之间的季节性变化或照明变化造成的。数据扩增是一个解决这一问题的有吸引力的方法,因为它不需要提供额外的数据。然而,现有的增强方法盲目地干扰所有像素,因此无法取得令人满意的性能。为解决这一问题,我们建议RADA,这个系统的目的是集中研究图像的几何信息部分。结果,它学会产生最小的图像扰动,仍然能够将网络弄乱。我们表明,当这些例子被用作增强时,它会大大改进稳健性。我们表明,我们的方法比SOTA本地化模型(例如,AtLoc和MapNet)的精度高两倍于在“无法预见的天气条件”测试时,我们的方法比SOTA的本地化模型(例如,Atloc和MetNet)的精度高一倍。