PointGoal navigation in indoor environment is a fundamental task for personal robots to navigate to a specified point. Recent studies solved this PointGoal navigation task with near-perfect success rate in photo-realistically simulated environments, under the assumptions with noiseless actuation and most importantly, perfect localization with GPS and compass sensors. However, accurate GPS signalis difficult to be obtained in real indoor environment. To improve the PointGoal navigation accuracy without GPS signal, we use visual odometry (VO) and propose a novel action integration module (AIM) trained in unsupervised manner. Sepecifically, unsupervised VO computes the relative pose of the agent from the re-projection error of two adjacent frames, and then replaces the accurate GPS signal with the path integration. The pseudo position estimated by VO is used to train action integration which assists agent to update their internal perception of location and helps improve the success rate of navigation. The training and inference process only use RGB, depth, collision as well as self-action information. The experiments show that the proposed system achieves satisfactory results and outperforms the partially supervised learning algorithms on the popular Gibson dataset.
翻译:在室内环境中进行点目标导航是个人机器人导航到指定点的基本任务。最近的研究在近乎完美的成功率下解决了这个点目标导航任务,这是在假设无噪声执行以及最重要的是GPS和罗盘传感器的完美定位的条件下完成的。然而,真实室内环境中很难获得精确的GPS信号。为了在没有GPS信号的情况下提高点目标导航精度,我们使用视觉里程计(VO)并提出了一种新的以无监督的方式训练的动作整合模块(AIM)。具体而言,无监督的VO通过两个相邻帧的重投影误差计算代理相对姿态,然后使用路径整合替代精确的GPS信号。VO估计的伪位置用于训练动作整合,该整合帮助代理更新其位置的内部感知,并有助于提高导航成功率。训练和推理过程只使用RGB、深度、碰撞以及自我行为信息。实验表明,所提出的系统实现了满意的结果,并在流行的Gibson数据集上优于部分监督学习算法。