Several approaches have been developed to answer specific questions that a user may have about an AI system that can plan and act. However, the problems of identifying which questions to ask and that of computing a user-interpretable symbolic description of the overall capabilities of the system have remained largely unaddressed. This paper presents an approach for addressing these problems by learning user-interpretable symbolic descriptions of the limits and capabilities of a black-box AI system using low-level simulators. It uses a hierarchical active querying paradigm to generate questions and to learn a user-interpretable model of the AI system based on its responses. In contrast to prior work, we consider settings where imprecision of the user's conceptual vocabulary precludes a direct expression of the agent's capabilities. Furthermore, our approach does not require assumptions about the internal design of the target AI system or about the methods that it may use to compute or learn task solutions. Empirical evaluation on several game-based simulator domains shows that this approach can efficiently learn symbolic models of AI systems that use a deterministic black-box policy in fully observable scenarios.
翻译:开发了几种方法来回答用户可能拥有的关于能够规划和采取行动的AI系统的具体问题,然而,在确定应问的问题和计算用户解释的关于该系统总体能力的象征性描述方面,问题基本上仍未解决,本文件介绍了一种解决这些问题的方法,即学习用户解释的关于使用低级别模拟器的黑盒AI系统的局限性和能力的象征性描述,它使用一个等级积极的查询模式来产生问题,并学习基于其回应的AI系统的用户解释模型。与先前的工作相比,我们考虑了用户概念词汇不精确而无法直接表达代理人能力的情况。此外,我们的方法并不要求假设目标AI系统的内部设计或它可能用来计算或学习任务解决方案的方法。对若干基于游戏的模拟领域的“经验性评估”表明,这种方法可以有效地学习在完全可观察的情景中采用确定型黑盒政策的AI系统的象征性模型。