As one of the prevalent methods to achieve automation systems, Imitation Learning (IL) presents a promising performance in a wide range of domains. However, despite the considerable improvement in policy performance, the corresponding research on the explainability of IL models is still limited. Inspired by the recent approaches in explainable artificial intelligence methods, we proposed a model-agnostic explaining framework for IL models called R2RISE. R2RISE aims to explain the overall policy performance with respect to the frames in demonstrations. It iteratively retrains the black-box IL model from the randomized masked demonstrations and uses the conventional evaluation outcome environment returns as the coefficient to build an importance map. We also conducted experiments to investigate three major questions concerning frames' importance equality, the effectiveness of the importance map, and connections between importance maps from different IL models. The result shows that R2RISE successfully distinguishes important frames from the demonstrations.
翻译:作为实现自动化系统的主要方法之一,模拟学习(IL)在广泛的领域展现了有希望的业绩,然而,尽管政策业绩有了相当大的改进,关于IL模型可解释性的相应研究仍然有限,由于最近在可解释的人工智能方法方面采用的方法,我们为名为R2RISE的IL模型提出了一个模型――不可知性解释框架。R2RISE旨在解释演示框架的总体政策业绩。它反复反复将黑盒 IL模型从随机化蒙面演示中重新引入黑盒 IL模型,并利用常规评价结果环境回报作为建立重要地图的系数。我们还进行了实验,以调查三个主要问题:框架的重要性平等、重要地图的有效性以及不同IL模型的重要性地图之间的联系。结果显示,R2RISE成功地将重要的框架与演示相区别开来。