Robot navigation in dynamic environments shared with humans is an important but challenging task, which suffers from performance deterioration as the crowd grows. In this paper, multi-subgoal robot navigation approach based on deep reinforcement learning is proposed, which can reason about more comprehensive relationships among all agents (robot and humans). Specifically, the next position point is planned for the robot by introducing history information and interactions in our work. Firstly, based on subgraph network, the history information of all agents is aggregated before encoding interactions through a graph neural network, so as to improve the ability of the robot to anticipate the future scenarios implicitly. Further consideration, in order to reduce the probability of unreliable next position points, the selection module is designed after policy network in the reinforcement learning framework. In addition, the next position point generated from the selection module satisfied the task requirements better than that obtained directly from the policy network. The experiments demonstrate that our approach outperforms state-of-the-art approaches in terms of both success rate and collision rate, especially in crowded human environments.
翻译:与人类共享的动态环境中的机器人导航是一项重要但具有挑战性的任务,随着人群的增长,其性能会恶化。在本文中,提出了基于深层强化学习的多子目标机器人导航方法,这可以说明所有代理商(机器人和人类)之间更全面的关系。具体地说,通过在我们的工作中引入历史信息和互动,为机器人规划下一个位置点。首先,根据子图网络,所有代理商的历史信息在通过图形神经网络对互动进行编码之前会汇总,以便提高机器人对未来情景进行隐含预测的能力。为了降低下一个定位点不可靠的可能性,在强化学习框架中,选择模块是在政策网络之后设计的。此外,从选择模块产生的下一个位置比直接从政策网络获得的更符合任务要求。实验表明,我们的方法在成功率和碰撞率方面都超越了最先进的方法,特别是在拥挤的人类环境中。