Interpretable machine learning (IML) is concerned with the behavior and the properties of machine learning models. Scientists, however, are only interested in the model as a gateway to understanding the modeled phenomenon. We show how to develop IML methods such that they allow insight into relevant phenomenon properties. We argue that current IML research conflates two goals of model-analysis -- model audit and scientific inference. Thereby, it remains unclear if model interpretations have corresponding phenomenon interpretation. Building on statistical decision theory, we show that ML model analysis allows to describe relevant aspects of the joint data probability distribution. We provide a five-step framework for constructing IML descriptors that can help in addressing scientific questions, including a natural way to quantify epistemic uncertainty. Our phenomenon-centric approach to IML in science clarifies: the opportunities and limitations of IML for inference; that conditional not marginal sampling is required; and, the conditions under which we can trust IML methods.
翻译:解释机器学习(IML)与机器学习模型的行为和性质有关。然而,科学家们只对模型感兴趣,认为模型是了解模型现象的网关。我们展示了如何开发IML方法,以便能够洞察到相关现象的特性。我们争辩说,目前的IML研究将模型分析的两个目标 -- -- 示范审计和科学推理 -- -- 结合起来。因此,模型解释是否具有相应的现象解释仍然不清楚。根据统计决策理论,我们表明ML模型分析能够描述联合数据概率分布的相关方面。我们为建立IML描述符提供了五步框架,有助于解决科学问题,包括用自然方式量化特征不确定性。我们在科学中以现象为中心的IML方法澄清了:IML的推断机会和局限性;需要有条件的、不边缘的取样;以及我们可以信任IML方法的条件。