While Explainable Artificial Intelligence (XAI) is increasingly expanding more areas of application, little has been applied to make deep Reinforcement Learning (RL) more comprehensible. As RL becomes ubiquitous and used in critical and general public applications, it is essential to develop methods that make it better understood and more interpretable. This study proposes a novel approach to explain cooperative strategies in multiagent RL using Shapley values, a game theory concept used in XAI that successfully explains the rationale behind decisions taken by Machine Learning algorithms. Through testing common assumptions of this technique in two cooperation-centered socially challenging multi-agent environments environments, this article argues that Shapley values are a pertinent way to evaluate the contribution of players in a cooperative multi-agent RL context. To palliate the high overhead of this method, Shapley values are approximated using Monte Carlo sampling. Experimental results on Multiagent Particle and Sequential Social Dilemmas show that Shapley values succeed at estimating the contribution of each agent. These results could have implications that go beyond games in economics, (e.g., for non-discriminatory decision making, ethical and responsible AI-derived decisions or policy making under fairness constraints). They also expose how Shapley values only give general explanations about a model and cannot explain a single run, episode nor justify precise actions taken by agents. Future work should focus on addressing these critical aspects.
翻译:虽然可解释的人工智能(XAI)正在日益扩大更多的应用领域,但很少应用这种方法来使深强化学习(RL)更加容易理解。由于RL变得无处不在,并被用于关键和一般的公共应用中,因此有必要制定方法,使其更易理解和解释。本研究报告提出了一种新颖的方法,用以解释多试剂RL的合作战略,使用Shapley价值,这是XAI使用的游戏理论概念,成功地解释了机器学习算法所作决定背后的理由。通过在两个以合作为中心、具有社会挑战性的多代理人环境环境中测试这一技术的共同假设,这一文章认为,Spley价值是评价参与者在合作性多代理人RL背景下所作贡献的恰当方法。要缓和这一方法的高间接意义,使用Monte Carlo抽样比较了Sample值。多剂粒子和顺序社会变迁概念的实验结果表明,Shaly价值在估计每个代理人的贡献方面是成功的。这些结果可能产生的影响超越了经济领域的游戏,(例如,为了无歧视性的、道德和负责任的、也不可能作出单一的决定,因此作出明确的解释。