Machine learning becomes increasingly important to control the behavior of safety and financially critical components in sophisticated environments, where the inability to understand learned components in general, and neural nets in particular, poses serious obstacles to their adoption. Explainability and interpretability methods for learned systems have gained considerable academic attention, but the focus of current approaches on only one aspect of explanation, at a fixed level of abstraction, and limited if any formal guarantees, prevents those explanations from being digestible by the relevant stakeholders (e.g., end users, certification authorities, engineers) with their diverse backgrounds and situation-specific needs. We introduce Fanoos, a flexible framework for combining formal verification techniques, heuristic search, and user interaction to explore explanations at the desired level of granularity and fidelity. We demonstrate the ability of Fanoos to produce and adjust the abstractness of explanations in response to user requests on a learned controller for an inverted double pendulum and on a learned CPU usage model.
翻译:机器学习对于控制复杂环境中安全和财政关键组成部分的行为越来越重要,因为在复杂环境中,无法理解一般学到的部件,特别是神经网,对采用这些部件构成严重障碍。学习系统的可解释性和可解释性方法在学术上引起了相当大的关注,但目前方法只侧重于一个解释方面,即固定的抽象程度,如果正式保障有限,则这些解释不能被具有不同背景和具体国情需要的相关利益攸关方(如最终用户、验证局、工程师)消化。我们引入了Fanoos,这是一个灵活的框架,将正式核查技术、超常搜索和用户互动结合起来,以探讨在预期的颗粒度和忠诚度水平上的解释。我们表明Fanoos有能力编制和调整解释的抽象性,以回应用户对一个有知识的控制器要求,即一个反向的双顶和学习的CPU使用模式。