As input data distributions evolve, the predictive performance of machine learning models tends to deteriorate. In the past, predictive performance was considered the key indicator to monitor. However, explanation aspects have come to attention within the last years. In this work, we investigate how model predictive performance and model explanation characteristics are affected under distribution shifts and how these key indicators are related to each other for tabular data. We find that the modeling of explanation shifts can be a better indicator for the detection of predictive performance changes than state-of-the-art techniques based on representations of distribution shifts. We provide a mathematical analysis of different types of distribution shifts as well as synthetic experimental examples.
翻译:随着投入数据分布的演变,机器学习模型的预测性能往往会恶化。过去,预测性能被视为监测的关键指标。然而,过去几年中,解释方面引起了注意。在这项工作中,我们调查了模型预测性能和模型解释特性在分配变化中如何受到影响,以及这些关键指标在表格数据中如何相互关联。我们发现,解释性变化的模型性能变化的检测指标可以比基于分配变化表现的最新技术更好地。我们提供了对不同类型分配性变化的数学分析以及合成实验实例。