Many existing approaches for estimating feature importance are problematic because they ignore or hide dependencies among features. A causal graph, which encodes the relationships among input variables, can aid in assigning feature importance. However, current approaches that assign credit to nodes in the causal graph fail to explain the entire graph. In light of these limitations, we propose Shapley Flow, a novel approach to interpreting machine learning models. It considers the entire causal graph, and assigns credit to \textit{edges} instead of treating nodes as the fundamental unit of credit assignment. Shapley Flow is the unique solution to a generalization of the Shapley value axioms to directed acyclic graphs. We demonstrate the benefit of using Shapley Flow to reason about the impact of a model's input on its output. In addition to maintaining insights from existing approaches, Shapley Flow extends the flat, set-based, view prevalent in game theory based explanation methods to a deeper, \textit{graph-based}, view. This graph-based view enables users to understand the flow of importance through a system, and reason about potential interventions.
翻译:许多现有的估计特征重要性的方法都存在问题,因为它们忽视或隐藏了各种特征之间的依赖性。一个记录输入变量之间关系的因果图表可以帮助确定特性的重要性。然而,目前对因果图表中节点的信用分配方法未能解释整个图表。鉴于这些局限性,我们建议采用沙普利流这一解释机器学习模型的新颖方法。它考虑到整个因果图表,并将偏爱节点作为信用分配的基本单位给予\ textit{sedge}。沙普利流是将沙普利值的正值普遍化为引导循环图的独特解决办法。我们展示了使用沙普利流动的好处,以说明模型投入对其产出的影响。除了保持现有方法的洞察外,沙普利流将基于游戏理论的简单、基于固定、普遍观点的解释方法扩展为更深层次的、 textitit{graphy} 观点。这种基于图表的观点使用户能够了解通过系统和潜在干预的理由来理解重要性的流动。