This work presents a decentralized motion planning framework for addressing the task of multi-robot navigation using deep reinforcement learning. A custom simulator was developed in order to experimentally investigate the navigation problem of 4 cooperative non-holonomic robots sharing limited state information with each other in 3 different settings. The notion of decentralized motion planning with common and shared policy learning was adopted, which allowed robust training and testing of this approach in a stochastic environment since the agents were mutually independent and exhibited asynchronous motion behavior. The task was further aggravated by providing the agents with a sparse observation space and requiring them to generate continuous action commands so as to efficiently, yet safely navigate to their respective goal locations, while avoiding collisions with other dynamic peers and static obstacles at all times. The experimental results are reported in terms of quantitative measures and qualitative remarks for both training and deployment phases.
翻译:这项工作提出了利用深层强化学习解决多机器人导航任务的分散化动议规划框架,开发了一个定制模拟器,以实验性地调查在三个不同环境中相互交流有限国家信息的4个合作型非超光层机器人的导航问题,采用了分散化动议规划的概念,并共同和共享政策学习,从而得以在随机环境中对这种方法进行强有力的培训和测试,因为代理商是相互独立的,表现出不同步的运动行为。由于向代理商提供稀少的观测空间,要求他们产生连续行动指令,以便高效、安全地前往各自的目标地点,同时避免与其他动态同行发生碰撞,避免随时出现固定障碍,实验结果在培训和部署阶段的定量措施和定性说明方面都有报告。