Reinforcement Learning approaches are becoming increasingly popular in various key disciplines, including robotics and healthcare. However, many of these systems are complex and non-interpretable, making it challenging for non-AI experts to understand or intervene in their decisions. One of the challenges of explaining RL agent behavior is that, when learning to predict future expected rewards, agents discard contextual information about their experiences when training in an environment and rely solely on expected utility. We propose a technique, Experiential Explanations, for generating local counterfactual explanations that can answer users' why-not questions by explaining the qualitative effects of the various environmental rewards on the agent's behavior. We achieve this by training additional models alongside the policy model. These models, called influence predictors, capture how different reward sources influence the agent's policy, thus restoring lost contextual information about how the policy reflects the environment. To generate explanations, we use these influence predictors in addition to the policy model to contrast between the agent's intended behavior trajectory and a counterfactual trajectory suggested by the user. A human evaluation study revealed that participants had a higher probability of correctly predicting the agent's subsequent action when presented with Experiential Explanations than other explanation types. Moreover, compared to other baseline types, participants found Experiential Explanations more useful and more often utilized the kinds of information presented in them when reasoning about the agent's actions. Experiential Explanations also outperformed other explanations in understandability, satisfaction, amount of details, completeness, usefulness, and accuracy.
翻译:在包括机器人和医疗保健在内的各种关键学科中,强化学习方法越来越受欢迎。然而,许多这些系统是复杂和不可解释的,使得非AI专家难以理解或干预其决定。解释RL代理行为的挑战之一是,当学会预测未来预期的回报时,代理商抛弃了有关其经验的背景资料,而完全依赖预期效用。我们建议一种技术,即“经验解释”,通过解释各种环境奖赏对代理商行为的质量影响来回答用户的错误原因。我们通过培训更多的模型来实现这一目标。这些模型称为影响力预测者,捕捉不同奖赏来源如何影响代理商的政策,从而恢复关于政策如何反映环境的丧失的背景资料。为了作出解释,我们除了政策模型之外,还使用这些影响预测者来比较该代理商的预期行为轨迹和用户建议的反事实轨迹。一项人类评估研究表明,参与者在解释代理人随后的行动时,更准确地预测其准确性的可能性更高,在与解释性解释性定义的其他类型中,还经常使用其他解释性解释类型时,比解释性解释性的其他解释类别更准确性。