Cooperative multi-robot teams need to be able to explore cluttered and unstructured environments together while dealing with communication challenges. Specifically, during communication dropout, local information about robots can no longer be exchanged to maintain robot team coordination. Therefore, robots need to consider high-level teammate intentions during action selection. In this paper, we present the first Macro Action Decentralized Exploration Network (MADE-Net) using multi-agent deep reinforcement learning to address the challenges of communication dropouts during multi-robot exploration in unseen, unstructured, and cluttered environments. Simulated robot team exploration experiments were conducted and compared to classical and deep reinforcement learning methods. The results showed that our MADE-Net method was able to outperform all benchmark methods in terms of computation time, total travel distance, number of local interactions between robots, and exploration rate across various degrees of communication dropouts; highlighting the effectiveness and robustness of our method.
翻译:合作型多机器人团队在应对通信挑战时,需要能够共同探索杂乱无章的环境。具体地说,在通信中断期间,关于机器人的当地信息无法再为维持机器人团队协调而交流。因此,机器人在选择行动时需要考虑高层次团队意向。在本文中,我们介绍了第一个宏观行动分散探索网络(MADE-Net),使用多试剂深度强化学习来解决在不可见、无结构的和封闭式环境中进行多机器人探索期间通信中断的挑战。进行了模拟机器人团队探索实验,并与经典和深层强化学习方法进行了比较。结果显示,我们的软件网络方法在计算时间、总旅行距离、机器人之间的本地互动次数以及不同程度通信辍学的探索率等方面都超过了所有基准方法;强调了我们方法的有效性和稳健性。