This paper surveys the field of deep multiagent reinforcement learning. The combination of deep neural networks with reinforcement learning has gained increased traction in recent years and is slowly shifting the focus from single-agent to multiagent environments. Dealing with multiple agents is inherently more complex as (a) the future rewards depend on multiple players' joint actions and (b) the computational complexity increases. We present the most common multiagent problem representations and their main challenges, and identify five research areas that address one or more of these challenges: centralised training and decentralised execution, opponent modelling, communication, efficient coordination, and reward shaping. We find that many computational studies rely on unrealistic assumptions or are not generalisable to other settings; they struggle to overcome the curse of dimensionality or nonstationarity. Approaches from psychology and sociology capture promising relevant behaviours, such as communication and coordination, to help agents achieve better performance in multiagent settings. We suggest that, for multiagent reinforcement learning to be successful, future research should address these challenges with an interdisciplinary approach to open up new possibilities in multiagent reinforcement learning.
翻译:本文调查了深多试剂强化学习领域。深神经网络与强化学习的结合近年来获得了更大的牵引力,正在慢慢地将重点从单一试剂环境转移到多试剂环境。处理多剂环境本身就更为复杂,因为(a) 未来的奖励取决于多个行为者的联合行动,和(b) 计算复杂性的增加。我们提出了最常见的多剂问题表述及其主要挑战,并确定了应对其中一项或多项挑战的五个研究领域:集中培训和分散执行、对手建模、沟通、高效协调和奖赏塑造。我们发现,许多计算研究依赖不切实际的假设,或者无法向其他环境推广;它们努力克服维度或非常态的诅咒。心理学和社会学的方法抓住了有希望的相关行为,例如沟通和协调,以帮助多剂机构在多剂环境中取得更好的表现。我们建议,为了多剂强化学习取得成功,今后的研究应当应对这些挑战,采取跨学科方法,为多剂强化学习开辟新的可能性。