Explainability of a classification model is crucial when deployed in real-world decision support systems. Explanations make predictions actionable to the user and should inform about the capabilities and limitations of the system. Existing explanation methods, however, typically only provide explanations for individual predictions. Information about conditions under which the classifier is able to support the decision maker is not available, while for instance information about when the system is not able to differentiate classes can be very helpful. In the development phase it can support the search for new features or combining models, and in the operational phase it supports decision makers in deciding e.g. not to use the system. This paper presents a method to explain the qualities of a trained base classifier, called PERFormance EXplainer (PERFEX). Our method consists of a meta tree learning algorithm that is able to predict and explain under which conditions the base classifier has a high or low error or any other classification performance metric. We evaluate PERFEX using several classifiers and datasets, including a case study with urban mobility data. It turns out that PERFEX typically has high meta prediction performance even if the base classifier is hardly able to differentiate classes, while giving compact performance explanations.
翻译:当在现实世界的决策支持系统中部署时,分类模型的可解释性至关重要。解释使预测对用户具有可操作性,并应当告知系统的能力和局限性。但现有的解释方法通常只为个别预测提供解释。分类者能够支持决策者的条件信息不详,而系统无法区分类别的信息则非常有用。在开发阶段,系统可以支持寻找新的特征或组合模型,在操作阶段,它可以支持决策者决定不使用系统等。本文提出了一个解释受过训练的基础分类者质量的方法,称为Pperforance Explainer(PERFEX)。我们的方法包括一种元树学习算法,能够预测和解释基础分类者在哪些条件下存在高或低误差,或任何其他分类性能衡量标准。我们利用若干分类者和数据集,包括城市流动性数据的案例研究,来评价PERFEX。它通常具有很高的元预测性能,即使基础分类者很难区分等级,同时提供压缩的绩效解释。