Solving hard-exploration environments in an important challenge in Reinforcement Learning. Several approaches have been proposed and studied, such as Intrinsic Motivation, co-evolution of agents and tasks, and multi-agent competition. In particular, the interplay between multiple agents has proven to be capable of generating human-relevant emergent behaviour that would be difficult or impossible to learn in single-agent settings. In this work, an extensible competitive environment for multi-agent interplay was developed, which features realistic physics and human-relevant semantics. Moreover, several experiments on different variants of this environment were performed, resulting in some simple emergent strategies and concrete directions for future improvement. The content presented here is part of the author's thesis "Multi-Agent Interplay in a Competitive Survival Environment" for the Master's Degree in Artificial Intelligence and Robotics at Sapienza University of Rome, 2022.
翻译:在强化学习的重大挑战中解决硬探索环境; 提出和研究了若干办法,例如自然动力、代理人和任务的共同演变以及多代理人竞争等; 特别是,多个代理人之间的相互作用证明能够产生在单一代理人环境中难以或不可能学习的与人类有关的突发行为; 在这项工作中,为多代理人相互作用发展了一种可扩展的竞争环境,这种互动具有现实的物理学和与人类有关的语义学特征; 此外,还就这种环境的不同变种进行了若干试验,结果产生了一些简单的新兴战略和未来的具体改进方向; 这里介绍的内容是作者为2022年罗马Sapienza大学人工智能和机器人学硕士学位编写的论文“竞争性生存环境中的多代理人互动”的一部分。