Human explanation (e.g., in terms of feature importance) has been recently used to extend the communication channel between human and agent in interactive machine learning. Under this setting, human trainers provide not only the ground truth but also some form of explanation. However, this kind of human guidance was only investigated in supervised learning tasks, and it remains unclear how to best incorporate this type of human knowledge into deep reinforcement learning. In this paper, we present the first study of using human visual explanations in human-in-the-loop reinforcement learning (HRL). We focus on the task of learning from feedback, in which the human trainer not only gives binary evaluative "good" or "bad" feedback for queried state-action pairs, but also provides a visual explanation by annotating relevant features in images. We propose EXPAND (EXPlanation AugmeNted feeDback) to encourage the model to encode task-relevant features through a context-aware data augmentation that only perturbs irrelevant features in human salient information. We choose five tasks, namely Pixel-Taxi and four Atari games, to evaluate the performance and sample efficiency of this approach. We show that our method significantly outperforms methods leveraging human explanation that are adapted from supervised learning, and Human-in-the-loop RL baselines that only utilize evaluative feedback.
翻译:人类的解释(例如,在特征重要性方面)最近被用来扩大人与代理人之间互动机器学习的沟通渠道。在这个背景下,人类教官不仅提供地面真相,而且提供某种形式的解释。然而,这种人类指导只是通过监督的学习任务来调查,而且仍然不清楚如何最好地将这种人类知识纳入深入的强化学习中。在本文件中,我们介绍关于将人类视觉解释用于加强人与人之间在流动中的强化学习(HRL)的首次研究。我们侧重于从反馈中学习的任务,在反馈中,人类教官不仅为询问的国家行动配对提供二元评价“良好”或“坏”反馈,而且还通过在图像中说明相关特征提供直观解释。我们建议EXPAND(EX Planation AgmeNed flobDback)鼓励模型将任务相关特征纳入深层次强化数据,而该数据增强只能渗透到人类突出信息中的不相干的特点。我们选择了五项任务,即Pixel-Taxi和四部Atari游戏,来评估业绩和抽样反馈效率,我们从这一基准中大大利用了人类评估方法。