Recent works have applied the Proximal Policy Optimization (PPO) to the multi-agent cooperative tasks, such as Independent PPO (IPPO); and vanilla Multi-agent PPO (MAPPO) which has a centralized value function. However, previous literature shows that MAPPO may not perform as well as Independent PPO (IPPO) and the Fine-tuned QMIX on Starcraft Multi-Agent Challenge (SMAC). MAPPO-Feature-Pruned (MAPPO-FP) improves the performance of MAPPO by the carefully designed agent-specific features, which is is not friendly to algorithmic utility. By contrast, we find that MAPPO faces the problem of \textit{The Policies Overfitting in Multi-agent Cooperation(POMAC)}, as they learn policies by the sampled shared advantage values. Then POMAC may lead to updating the multi-agent policies in a suboptimal direction and prevent the agents from exploring better trajectories. In this paper, to mitigate the multi-agent policies overfitting, we propose a novel policy perturbation method, which disturb the advantage values via random Gaussian noise. The experimental results show that our method outperforms the Fine-tuned QMIX, MAPPO-FP, and achieves SOTA on SMAC without agent-specific features. We open-source the code at \url{https://github.com/hijkzzz/noisy-mappo}.
翻译:近期的著作应用了Proximal政策优化(PPO)来完成多试剂合作任务,如独立PPO(IPPO)和具有集中价值功能的香草多试PPO(MAPPO),然而,以前的文献表明,MAPPO可能不会像独立PPO(IPPO)和关于Starcal多点挑战(SMAC)的微调QMIX那样发挥作用。MAPPO-Fature-Pruned(MAPO-FP)通过精心设计的特制代理人特有特征来改进MAPPO的性能,这种特有特征对算工具的实用性不友好。相比之下,我们发现MAPO面临“Textit{在多点合作(POMAC)中过度适应政策”的问题,因为它们通过抽样的共享优势值来学习政策。 之后,PAPMAC可能会导致以亚优方向更新多点政策,防止代理商探索更好的公开轨迹。在本文中,我们建议一种创新的政策性政策,我们建议一种不透性政策,在Stourbil-MICFAFI 方法上,通过随机分析方法来展示优势。