Planning coverage path for multiple robots in a decentralized way enhances robustness to coverage tasks handling uncertain malfunctions. To achieve high efficiency in a distributed manner for each single robot, a comprehensive understanding of both the complicated environments and cooperative agents intent is crucial. Unfortunately, existing works commonly consider only part of these factors, resulting in imbalanced subareas or unnecessary overlaps. To tackle this issue, we introduce a Decentralized reinforcement learning framework with dual guidance to train each agent to solve the decentralized multiple coverage path planning problem straightly through the environment states. As distributed robots require others intentions to perform better coverage efficiency, we utilize two guidance methods, artificial potential fields and heuristic guidance, to include and integrate others intentions into observations for each robot. With our constructed framework, results have shown our agents successfully learn to determine their own subareas while achieving full coverage, balanced subareas and low overlap rates. We then implement spanning tree cover within those subareas to construct actual routes for each robot and complete given coverage tasks. Our performance is also compared with the state of the art decentralized method showing at most 10 percent lower overlap rates while performing high efficiency in similar environments.
翻译:以分散方式规划多机器人的覆盖范围路径,可以增强稳健性,以覆盖处理不确定故障的任务。为了以分布方式实现每个机器人的高效,全面理解复杂的环境和合作剂的意图至关重要。 不幸的是,现有工作通常只考虑这些因素的一部分,导致子领域不平衡或不必要的重叠。为了解决这一问题,我们引入了一个分散强化学习框架,同时提供双重指导,培训每个代理人员直接通过环境国家解决分散的多覆盖路径规划问题。分布式机器人要求其他人打算提高覆盖效率,因此我们使用两种指导方法,即人工潜在字段和超常指导,将其他意图纳入和纳入对每个机器人的观察中。根据我们构建的框架,结果显示我们的代理人员成功地学会了确定自己的子领域,同时实现了全面覆盖、平衡的子领域和低的重叠率。然后,我们在这些子领域实施覆盖树覆盖框架,以便为每个机器人建造实际路径并完成覆盖任务。我们的业绩与分散式方法的状态相比,我们的业绩也显示在类似环境中执行高效率的同时,最低10%的重叠率。