In multi-agent reinforcement learning, the behaviors that agents learn in a single Markov Game (MG) are typically confined to the given agent number (i.e., population size). Every single MG induced by varying population sizes may possess distinct optimal joint strategies and game-specific knowledge, which are modeled independently in modern multi-agent algorithms. In this work, we focus on creating agents that generalize across population-varying MGs. Instead of learning a unimodal policy, each agent learns a policy set that is formed by effective strategies across a variety of games. We propose Meta Representations for Agents (MRA) that explicitly models the game-common and game-specific strategic knowledge. By representing the policy sets with multi-modal latent policies, the common strategic knowledge and diverse strategic modes are discovered with an iterative optimization procedure. We prove that as an approximation to a constrained mutual information maximization objective, the learned policies can reach Nash Equilibrium in every evaluation MG under the assumption of Lipschitz game on a sufficiently large latent space. When deploying it at practical latent models with limited size, fast adaptation can be achieved by leveraging the first-order gradient information. Extensive experiments show the effectiveness of MRA on both training performance and generalization ability in hard and unseen games.
翻译:在多试剂强化学习中,代理商在单一马可夫游戏(MG)中学习的行为通常限于特定代理商数目(即人口规模)。每个由不同人口规模诱导的单一MG都可能拥有独特的最佳联合战略和具体游戏知识,这些战略和知识以现代多试算法独立建模。在这项工作中,我们侧重于创建跨人口变化的MGs的通用代理商。每个代理商不学习单一模式政策,而是学习由各种游戏的有效战略构成的一套政策。我们提议为代理商提供明确模拟游戏常见和游戏特定战略知识的MMAS代表(MRA),通过代表具有多模式潜伏政策的政策组合,共同的战略知识和不同战略模式以迭代优化程序被发现。我们证明,作为对受限制的相互信息最大化目标的一种近似,在假定利普西茨游戏在足够大的潜在空间上形成有效战略,每个评价MGMS(Nash Equililiblicriumiumium)的每次评价中,每个评价都能够将精度应用规模有限的实际潜模型,在利用第一模型时,通过利用MMIS级能力测试和深层数据测试的硬度,从而显示硬度信息,可以实现。