With the rapid growth of machine learning, deep neural networks (DNNs) are now being used in numerous domains. Unfortunately, DNNs are "black-boxes", and cannot be interpreted by humans, which is a substantial concern in safety-critical systems. To mitigate this issue, researchers have begun working on explainable AI (XAI) methods, which can identify a subset of input features that are the cause of a DNN's decision for a given input. Most existing techniques are heuristic, and cannot guarantee the correctness of the explanation provided. In contrast, recent and exciting attempts have shown that formal methods can be used to generate provably correct explanations. Although these methods are sound, the computational complexity of the underlying verification problem limits their scalability; and the explanations they produce might sometimes be overly complex. Here, we propose a novel approach to tackle these limitations. We (1) suggest an efficient, verification-based method for finding minimal explanations, which constitute a provable approximation of the global, minimum explanation; (2) show how DNN verification can assist in calculating lower and upper bounds on the optimal explanation; (3) propose heuristics that significantly improve the scalability of the verification process; and (4) suggest the use of bundles, which allows us to arrive at more succinct and interpretable explanations. Our evaluation shows that our approach significantly outperforms state-of-the-art techniques, and produces explanations that are more useful to humans. We thus regard this work as a step toward leveraging verification technology in producing DNNs that are more reliable and comprehensible.
翻译:随着机器学习的迅速增长,深神经网络(DNNs)正在许多领域使用。 不幸的是,DNS是“黑箱”,不能被人类解释,这是安全临界系统中的一个重大关切。为了缓解这一问题,研究人员已开始研究可解释的AI(XAI)方法,该方法可以确定一组输入特征,这是DNN决定某项投入的原因。大多数现有技术都是杂乱无序的,无法保证所提供的解释的正确性。相比之下,最近令人振奋的尝试表明,正式方法可以用来产生可调适的正确解释。虽然这些方法是健全的,但基本核查问题的计算复杂性限制了其可伸缩性;为了缓解这一问题,他们提出的解释有时可能过于复杂。在这里,我们提出了一种新颖的方法来解决这些局限性。我们建议一种高效的、基于核查的方法,以找到最起码的解释,这是全球、最起码的精确的解释;(2)表明DNN核查如何帮助计算出更低和上层的最佳解释;(3) 提议采用更精准的计算方法,从而大大改进我们的精确性解释方法,从而大大改进我们的精确性解释。