To collaborate well with robots, we must be able to understand their decision making. Humans naturally infer other agents' beliefs and desires by reasoning about their observable behavior in a way that resembles inverse reinforcement learning (IRL). Thus, robots can convey their beliefs and desires by providing demonstrations that are informative for a human learner's IRL. An informative demonstration is one that differs strongly from the learner's expectations of what the robot will do given their current understanding of the robot's decision making. However, standard IRL does not model the learner's existing expectations, and thus cannot do this counterfactual reasoning. We propose to incorporate the learner's current understanding of the robot's decision making into our model of human IRL, so that a robot can select demonstrations that maximize the human's understanding. We also propose a novel measure for estimating the difficulty for a human to predict instances of a robot's behavior in unseen environments. A user study finds that our test difficulty measure correlates well with human performance and confidence. Interestingly, considering human beliefs and counterfactuals when selecting demonstrations decreases human performance on easy tests, but increases performance on difficult tests, providing insight on how to best utilize such models.
翻译:要与机器人合作,我们就必须能够理解他们的决策。人类自然地通过推理其他代理人的信仰和欲望,以类似于反强化学习(IRL)的方式推理他们的可见行为,自然地推断出其他代理人的信仰和欲望。因此,机器人可以通过为人类学习者的IRL提供信息的演示来表达他们的信仰和愿望。一个内容丰富的演示是一个与学习者对机器人在当前了解机器人决策的情况下对机器人将做什么的期望截然不同的新措施。然而,标准的IRL并不模拟学习者的现有期望,因此无法进行这一反事实推理。我们提议将学习者目前对机器人决策的理解纳入我们的人类IRL模型,以便机器人能够选择能够最大限度地提高人类理解度的演示。我们还提出了一个新措施,用以估计人类难以预测机器人在不可见环境中的行为实例。用户研究发现,我们的测试难度与人类业绩和信心密切相关。有趣的是,在选择示范时,考虑到人类信仰和反事实如何降低人类在简单测试中的性能,但提高人的性能。