Predictive process analytics focuses on predicting the future states of running instances of a business process. While advanced machine learning techniques have been used to increase accuracy of predictions, the resulting predictive models lack transparency. Current explainable machine learning methods, such as LIME and SHAP, can be used to interpret black box models. However, it is unclear how fit for purpose these methods are in explaining process predictive models. In this paper, we draw on evaluation measures used in the field of explainable AI and propose functionally-grounded evaluation metrics for assessing explainable methods in predictive process analytics. We apply the proposed metrics to evaluate the performance of LIME and SHAP in interpreting process predictive models built on XGBoost, which has been shown to be relatively accurate in process predictions. We conduct the evaluation using three open source, real-world event logs and analyse the evaluation results to derive insights. The research contributes to understanding the trustworthiness of explainable methods for predictive process analytics as a fundamental and key step towards human user-oriented evaluation.
翻译:预测过程分析方法侧重于预测商业过程的未来运行情况。虽然已经使用了先进的机器学习技术来提高预测的准确性,但由此产生的预测模型却缺乏透明度。目前可以解释的机器学习方法,如LIME和SHAP,可以用来解释黑盒模型;然而,尚不清楚这些方法在解释过程预测模型时是否适合目的。在本文件中,我们借鉴了在可解释的AI领域使用的评价措施,并提出了用于评估预测过程分析中可解释方法的基于功能的评价指标。我们运用了拟议的指标来评价LIME和SHAP在解释过程预测模型中的性能,这在过程预测中已经证明相对准确。我们用三种开放的源,即真实世界事件日志来进行评价,并分析评价结果以获得洞察力。研究有助于理解可解释的预测过程分析方法的可靠性,作为人类用户导向评价的基本和关键步骤。