The last decade witnessed an ever-increasing stream of successes in Machine Learning (ML). These successes offer clear evidence that ML is bound to become pervasive in a wide range of practical uses, including many that directly affect humans. Unfortunately, the operation of the most successful ML models is incomprehensible for human decision makers. As a result, the use of ML models, especially in high-risk and safety-critical settings is not without concern. In recent years, there have been efforts on devising approaches for explaining ML models. Most of these efforts have focused on so-called model-agnostic approaches. However, all model-agnostic and related approaches offer no guarantees of rigor, hence being referred to as non-formal. For example, such non-formal explanations can be consistent with different predictions, which renders them useless in practice. This paper overviews the ongoing research efforts on computing rigorous model-based explanations of ML models; these being referred to as formal explanations. These efforts encompass a variety of topics, that include the actual definitions of explanations, the characterization of the complexity of computing explanations, the currently best logical encodings for reasoning about different ML models, and also how to make explanations interpretable for human decision makers, among others.
翻译:过去十年中,机器学习(ML)取得了越来越多的成功,这些成功清楚地证明,ML必然在广泛的实际用途中变得普遍,包括许多直接影响到人类的用途。不幸的是,最成功的ML模型的运作对人类决策者来说是无法理解的。因此,使用ML模型,特别是在高风险和安全关键环境中,并非无所顾虑。近年来,一直在努力制定解释ML模型的方法。这些努力大多集中于所谓的模型-不可知性方法。然而,所有模型-不可知性和相关方法都无法保证严谨,因此被称为非正规。例如,这种非正式解释可能与不同的预测一致,使得这些预测在实践中毫无用处。本文概述了目前为计算基于模型的严格解释ML模型模型模型模型模型而作的研究工作;这些称为正式解释。这些努力包括各种专题,其中包括解释的实际定义、计算解释的复杂性、目前对不同ML模型进行推理的最符合逻辑的编码、对不同的ML模型作其他模型作出最合理的解释,以及人们如何作出最合理的解释。