Despite the recent advancement in multi-agent reinforcement learning (MARL), the MARL agents easily overfit the training environment and perform poorly in the evaluation scenarios where other agents behave differently. Obtaining generalizable policies for MARL agents is thus necessary but challenging mainly due to complex multi-agent interactions. In this work, we model the problem with Markov Games and propose a simple yet effective method, ranked policy memory (RPM), to collect diverse multi-agent trajectories for training MARL policies with good generalizability. The main idea of RPM is to maintain a look-up memory of policies. In particular, we try to acquire various levels of behaviors by saving policies via ranking the training episode return, i.e., the episode return of agents in the training environment; when an episode starts, the learning agent can then choose a policy from the RPM as the behavior policy. This innovative self-play training framework leverages agents' past policies and guarantees the diversity of multi-agent interaction in the training data. We implement RPM on top of MARL algorithms and conduct extensive experiments on Melting Pot. It has been demonstrated that RPM enables MARL agents to interact with unseen agents in multi-agent generalization evaluation scenarios and complete given tasks, and it significantly boosts the performance up to 402% on average.
翻译:尽管最近多试剂强化学习(MARL)取得了进展,但MARL代理商很容易地过度适应培训环境,在其他代理商行为不同的评估情景中表现不佳。因此,为MARL代理商制定普遍适用的政策是必要的,但主要由于复杂的多试剂互动关系而具有挑战性。在这项工作中,我们对Markov运动会的问题进行模拟,并提出一个简单而有效的方法,将政策记忆排序(RPM),以收集多种多试剂的轨迹,对MARL政策进行具有良好普遍性的培训。RPM的主要想法是保持对政策的回顾记忆。特别是,我们试图通过对培训事件回报进行排名来获得各种程度的保存政策,即代理商在培训环境中的复发;当事件开始时,学习代理商可以从RPM选择一项政策作为行为政策。这个创新的自我作用培训框架利用代理人过去的政策,保证培训数据中多试剂互动的多样性。我们实施RPMML算法,并在Melting Pot上进行广泛的实验。它已经完成了RPMMRP的升级,使MMER代理商能够大大地提升40个普通和普通的代理商对40级任务进行互动。