Existing approaches for the design of interpretable agent behavior consider different measures of interpretability in isolation. In this paper we posit that, in the design and deployment of human-aware agents in the real world, notions of interpretability are just some among many considerations; and the techniques developed in isolation lack two key properties to be useful when considered together: they need to be able to 1) deal with their mutually competing properties; and 2) an open world where the human is not just there to interpret behavior in one specific form. To this end, we consider three well-known instances of interpretable behavior studied in existing literature -- namely, explicability, legibility, and predictability -- and propose a revised model where all these behaviors can be meaningfully modeled together. We will highlight interesting consequences of this unified model and motivate, through results of a user study, why this revision is necessary.
翻译:设计可解释代理人行为的现有方法考虑孤立地解释的不同计量方法。在本文件中,我们假设,在现实世界中,人类觉悟代理人的设计和部署中,可解释性概念只是许多考虑中的一些;孤立地开发的技术缺乏在一起考虑时有用的两个关键属性:它们需要能够(1) 处理彼此竞争的特性;(2) 一个人类不仅存在以一种特定形式解释行为的开放世界。为此,我们考虑了现有文献中研究的三种众所周知的可解释行为实例 -- -- 即可解释性、可理解性和可预测性 -- -- 并提出一个修订模型,使所有这些行为都能一起有意义地建模。我们将强调这一统一模式的有趣后果,并通过用户研究结果激励人们,为什么有必要进行这一修订。