Global localisation from visual data is a challenging problem applicable to many robotics domains. Prior works have shown that neural networks can be trained to map images of an environment to absolute camera pose within that environment, learning an implicit neural mapping in the process. In this work we evaluate the applicability of such an approach to real-world robotics scenarios, demonstrating that by constraining the problem to 2-dimensions and significantly increasing the quantity of training data, a compact model capable of real-time inference on embedded platforms can be used to achieve localisation accuracy of several centimetres. We deploy our trained model onboard a UGV platform, demonstrating its effectiveness in a waypoint navigation task. Along with this work we will release a novel localisation dataset comprising simulated and real environments, each with training samples numbering in the tens of thousands.
翻译:从视觉数据中实现本地化是适用于许多机器人领域的一个具有挑战性的问题。先前的工程已经表明,神经网络可以接受培训,绘制环境图像图象,在环境中形成绝对摄像头,在此过程中学习隐含的神经制图。在这项工作中,我们评估了这种方法对现实世界机器人情景的可适用性,表明通过将问题限制为二层和大幅增加培训数据的数量,能够实时推断嵌入平台的紧凑模型可以用来实现几厘米的本地化精确度。我们在UGV平台上安装了我们经过训练的模型,展示了该模型的有效性,以某种方式显示其有效性。在这项工作中,我们将发布一个由模拟和真实环境组成的新的本地化数据集,每组培训样本数以万计。