Addressing the need for explainable Machine Learning has emerged as one of the most important research directions in modern Artificial Intelligence (AI). While the current dominant paradigm in the field is based on black-box models, typically in the form of (deep) neural networks, these models lack direct interpretability for human users, i.e., their outcomes (and, even more so, their inner working) are opaque and hard to understand. This is hindering the adoption of AI in safety-critical applications, where high interests are at stake. In these applications, explainable by design models, such as decision trees, may be more suitable, as they provide interpretability. Recent works have proposed the hybridization of decision trees and Reinforcement Learning, to combine the advantages of the two approaches. So far, however, these works have focused on the optimization of those hybrid models. Here, we apply MAP-Elites for diversifying hybrid models over a feature space that captures both the model complexity and its behavioral variability. We apply our method on two well-known control problems from the OpenAI Gym library, on which we discuss the "illumination" patterns projected by MAP-Elites, comparing its results against existing similar approaches.
翻译:在现代人工智能(AI)中,解决可解释的机器学习需要已成为最重要的研究方向之一。 目前这一领域的主导范式以黑盒模型为基础,通常以(深)神经网络为形式,但这些模型缺乏直接的人类用户可解释性,即其结果(特别是其内在工作)不透明,难以理解。这阻碍了在安全关键应用中采用AI,因为安全关键应用涉及重大利益。在这些应用中,由决策树等设计模型解释可能更合适,因为它们提供了解释性。最近的工作提出了决定树和强化学习的混合化,以结合这两种方法的优势。然而,这些模型侧重于这些混合模型的优化。在这里,我们应用MAP-Elites将混合模型多样化用于一个特征空间,既捕捉模型的复杂性,也捕捉其行为性变异性。我们在OpenAI Gym 图书馆的两个众所周知的控制问题上采用了我们的方法,我们讨论了“照明”模式与现有MAP-Els模拟结果的对比。