Multi-agent reinforcement learning (MARL) has been gaining extensive attention from academia and industries in the past few decades. One of the fundamental problems in MARL is how to evaluate different approaches comprehensively. Most existing MARL methods are evaluated in either video games or simplistic simulated scenarios. It remains unknown how these methods perform in real-world scenarios, especially multi-robot systems. This paper introduces a scalable emulation platform for multi-robot reinforcement learning (MRRL) called SMART to meet this need. Precisely, SMART consists of two components: 1) a simulation environment that provides a variety of complex interaction scenarios for training and 2) a real-world multi-robot system for realistic performance evaluation. Besides, SMART offers agent-environment APIs that are plug-and-play for algorithm implementation. To illustrate the practicality of our platform, we conduct a case study on the cooperative driving lane change scenario. Building off the case study, we summarize several unique challenges of MRRL, which are rarely considered previously. Finally, we open-source the simulation environments, associated benchmark tasks, and state-of-the-art baselines to encourage and empower MRRL research.
翻译:在过去几十年中,学术界和行业广泛关注多剂加固学习(MARL)问题。MARL的根本问题之一是如何全面评价不同方法。大多数现有的MARL方法都是在电子游戏或简单模拟情景中加以评价的。这些方法在现实世界情景中是如何运作的,特别是多机器人系统。本文介绍了一个可扩缩的多机器人加固学习模拟平台(MRRL),称为SMART,以满足这一需要。确切地说,SMART由两个部分组成:1)一个模拟环境,为培训提供各种复杂的互动情景;2)一个现实世界多机器人系统,用于现实的绩效评估。此外,SMART提供代理环境动画,作为算法执行的插头和功能。为了说明我们的平台的实用性,我们进行了一项关于合作驱动器变换情景的案例研究。根据案例研究,我们总结了MRRL的几项独特挑战,而以前很少考虑过这些挑战。最后,我们公开介绍了模拟环境、相关基准任务和状态的基线,以鼓励和赋予MRRRRRL的研究能力。