Reinforcement learning techniques achieved human-level performance in several tasks in the last decade. However, in recent years, the need for interpretability emerged: we want to be able to understand how a system works and the reasons behind its decisions. Not only we need interpretability to assess the safety of the produced systems, we also need it to extract knowledge about unknown problems. While some techniques that optimize decision trees for reinforcement learning do exist, they usually employ greedy algorithms or they do not exploit the rewards given by the environment. This means that these techniques may easily get stuck in local optima. In this work, we propose a novel approach to interpretable reinforcement learning that uses decision trees. We present a two-level optimization scheme that combines the advantages of evolutionary algorithms with the advantages of Q-learning. This way we decompose the problem into two sub-problems: the problem of finding a meaningful and useful decomposition of the state space, and the problem of associating an action to each state. We test the proposed method on three well-known reinforcement learning benchmarks, on which it results competitive with respect to the state-of-the-art in both performance and interpretability. Finally, we perform an ablation study that confirms that using the two-level optimization scheme gives a boost in performance in non-trivial environments with respect to a one-layer optimization technique.
翻译:在过去十年中,强化学习技术在几项任务中取得了人文层面的成绩。然而,近年来出现了解释性需要:我们希望能够理解一个系统如何运作,以及其决定背后的原因。我们不仅需要解释性来评估所生产的系统的安全性,我们还需要它来获取关于未知问题的知识。虽然有一些优化决策树来强化学习的技术确实存在,但它们通常采用贪婪的算法,或者它们不利用环境给予的奖励。这意味着这些技术很容易被困在本地的选取中。在这项工作中,我们提出了一种新颖的方法,用决定树来学习可解释的强化学习。我们提出了一个两级优化方案,将进化算法的优势与Q学习的优势结合起来。我们用这种方法将问题分解成两个子题:找到有意义和有用的州空间解析问题,以及将行动与各州联系起来的问题。我们用三种广为人知的强化学习基准测试了这三种方法,从而在使用决策树进行可解释性学习方面产生了竞争力。我们提出了一种两级优化技术的优势。最后,我们用一种方法将一个水平的绩效和解释性研究来确认一个层次的绩效。