The increasing number of regulations and expectations of predictive machine learning models, such as so called right to explanation, has led to a large number of methods promising greater interpretability. High demand has led to a widespread adoption of XAI techniques like Shapley values, Partial Dependence profiles or permutational variable importance. However, we still do not know enough about their properties and how they manifest in the context in which explanations are created by analysts, reviewed by auditors, and interpreted by various stakeholders. This paper highlights a blind spot which, although critical, is often overlooked when monitoring and auditing machine learning models: the effect of the reference data on the explanation calculation. We discuss that many model explanations depend directly or indirectly on the choice of the referenced data distribution. We showcase examples where small changes in the distribution lead to drastic changes in the explanations, such as a change in trend or, alarmingly, a conclusion. Consequently, we postulate that obtaining robust and useful explanations always requires supporting them with a broader context.
翻译:预测性机器学习模型(如所谓的解释权)的条例和期望越来越多,因此产生了大量方法,有可能提高解释性; 高需求导致广泛采用XAI技术,如Shapley值、部分依赖性剖面图或变异性重要性等; 然而,我们仍对其特性以及这些特性如何体现在分析家、审计员审查和各利益攸关方解释的解释中,了解得不够充分; 本文强调了一个盲点,虽然这是一个关键点,但在监测和审计机器学习模型中常常被忽视:参考数据对解释性计算的影响; 我们讨论许多示范解释直接或间接取决于引用数据分布的选择; 我们展示了一些实例,说明分配的微小变化导致解释发生急剧变化,例如趋势的变化,或令人震惊的结论。 因此,我们假设,获得有力和有用的解释总是需要在更广泛的范围内支持这些解释。