Although standard Machine Learning models are optimized for making predictions about observations, more and more they are used for making predictions about the results of actions. An important goal of Explainable Artificial Intelligence (XAI) is to compensate for this mismatch by offering explanations about the predictions of an ML-model which ensure that they are reliably action-guiding. As action-guiding explanations are causal explanations, the literature on this topic is starting to embrace insights from the literature on causal models. Here I take a step further down this path by formally defining the causal notions of sufficient explanations and counterfactual explanations. I show how these notions relate to (and improve upon) existing work, and motivate their adequacy by illustrating how different explanations are action-guiding under different circumstances. Moreover, this work is the first to offer a formal definition of actual causation that is founded entirely in action-guiding explanations. Although the definitions are motivated by a focus on XAI, the analysis of causal explanation and actual causation applies in general. I also touch upon the significance of this work for fairness in AI by showing how actual causation can be used to improve the idea of path-specific counterfactual fairness.
翻译:虽然标准的机器学习模型被优化用于预测观测结果,但越来越多的标准机器学习模型被用于预测行动结果,解释人工智能(XAI)的一个重要目标是通过解释ML模型的预测来弥补这种不匹配,这种预测确保了它们可靠的行动指导。由于行动指导解释是因果解释,关于这个专题的文献开始包含对因果模型文献的洞察力。我在这里迈出了一步,正式界定了充分解释和反事实解释的因果关系概念。我展示了这些概念与(并改进)现有工作的关系,并通过说明不同解释在不同情况下如何不同地指导行动来激励这些概念的充分性。此外,这项工作首先提供了完全基于行动指导解释的实际因果关系的正式定义。虽然定义的动机是侧重于XAI,但是对因果解释和实际因果关系的分析一般适用。我还谈到这项工作对于AI的公正性的意义,表明如何利用实际因果关系来改善特定路径的反事实公正性思想。