In this paper, we present an autonomous navigation system for goal-driven exploration of unknown environments through deep reinforcement learning (DRL). Points of interest (POI) for possible navigation directions are obtained from the environment and an optimal waypoint is selected, based on the available data. Following the waypoints, the robot is guided towards the global goal and the local optimum problem of reactive navigation is mitigated. Then, a motion policy for local navigation is learned through a DRL framework in a simulation. We develop a navigation system where this learned policy is integrated into a motion planning stack as the local navigation layer to move the robot between waypoints towards a global goal. The fully autonomous navigation is performed without any prior knowledge while a map is recorded as the robot moves through the environment. Experiments show that the proposed method has an advantage over similar exploration methods, without reliance on a map or prior information in complex static as well as dynamic environments.
翻译:在本文中,我们提出了一个自主导航系统,用于通过深层加固学习(DRL)对未知环境进行目标驱动的探索。从环境中获取可能的导航方向的感兴趣点,并根据现有数据选择最佳路径。在路标之后,机器人向全球目标方向前进,当地最佳反应导航问题得到缓解。然后,通过模拟的DRL框架学习当地导航运动政策。我们开发了一个导航系统,将这一已学习的政策作为当地导航层纳入运动规划堆放中,以将机器人移向全球目标。完全自主的导航是在没有任何事先知识的情况下进行的,而地图则记录为机器人穿越环境而记录。实验表明,拟议方法对类似的探索方法具有优势,而无需依靠地图或复杂静态和动态环境中的先前信息。