In explainable machine learning, local post-hoc explanation algorithms and inherently interpretable models are often seen as competing approaches. This work offers a partial reconciliation between the two by establishing a correspondence between Shapley Values and Generalized Additive Models (GAMs). We introduce $n$-Shapley Values, a parametric family of local post-hoc explanation algorithms that explain individual predictions with interaction terms up to order $n$. By varying the parameter $n$, we obtain a sequence of explanations that covers the entire range from Shapley Values up to a uniquely determined decomposition of the function we want to explain. The relationship between $n$-Shapley Values and this decomposition offers a functionally-grounded characterization of Shapley Values, which highlights their limitations. We then show that $n$-Shapley Values, as well as the Shapley Taylor- and Faith-Shap interaction indices, recover GAMs with interaction terms up to order $n$. This implies that the original Shapely Values recover GAMs without variable interactions. Taken together, our results provide a precise characterization of Shapley Values as they are being used in explainable machine learning. They also offer a principled interpretation of partial dependence plots of Shapley Values in terms of the underlying functional decomposition. A package for the estimation of different interaction indices is available at \url{https://github.com/tml-tuebingen/nshap}.
翻译:在可解释的机器学习中,本地的事后解释算法和固有的可解释模型往往被视为相互竞争的方法。这项工作通过在Shapley 值和通用Additive 模型(GAMs)之间建立对等关系,使两者之间部分调和。我们引入了美元-shaprey 值,这是一个对本地的事后解释算法的参数组合,解释个人预测与互动术语的互动值,最高达10美元。我们通过将参数美元差异,获得一系列解释,涵盖从Shaply 值一直到我们想要解释的功能独特确定的分解。美元-Shaply 值和这种分解式模型之间的关系提供了基于功能的对Shapley 值的描述,突出了它们的局限性。我们随后展示了美元- Shapley Taylor 和 Fuly-Shap 互动指数, 以互动术语恢复GAMs,最高达10美元。这意味着原始的Shapely 值恢复了GMS,而没有进行可变的交互作用。我们的成果还用了一个精确的模型来解释。