Trained AI systems and expert decision makers can make errors that are often difficult to identify and understand. Determining the root cause for these errors can improve future decisions. This work presents Generative Error Model (GEM), a generative model for inferring representational errors based on observations of an actor's behavior (either simulated agent, robot, or human). The model considers two sources of error: those that occur due to representational limitations -- "blind spots" -- and non-representational errors, such as those caused by noise in execution or systematic errors present in the actor's policy. Disambiguating these two error types allows for targeted refinement of the actor's policy (i.e., representational errors require perceptual augmentation, while other errors can be reduced through methods such as improved training or attention support). We present a Bayesian inference algorithm for GEM and evaluate its utility in recovering representational errors on multiple domains. Results show that our approach can recover blind spots of both reinforcement learning agents as well as human users.
翻译:经过培训的AI系统和专家决策者可以做出往往难以识别和理解的错误。 确定这些错误的根源可以改进未来的决策。 这项工作提出了“ 产生错误模型 ” ( GEM ), 这是根据对行为者行为( 模拟代理人、机器人或人类)的观察推断代表错误的遗传模型 。 模型考虑了两个错误来源: 由代表局限性( “ 盲点 ” ) 和无代表错误( 如执行过程中的噪音或行为者政策中存在的系统性错误)引起的错误。 区分这两个错误类型可以有针对性地完善行为者的政策( 即, 代表错误需要视觉增强, 而其他错误则可以通过改进培训或关注支持等方法减少 ) 。 我们为GEM 提出了一个贝耶斯的推断算法, 并评估其在恢复多个领域代表错误方面的效用。 结果显示,我们的方法可以回收强化学习代理人和人类用户的盲点 。