In many applications, it is important to be able to explain the decisions of machine learning systems. An increasingly popular approach has been to seek to provide \emph{counterfactual instance explanations}. These specify close possible worlds in which, contrary to the facts, a person receives their desired decision from the machine learning system. This paper will draw on literature from the philosophy of science to argue that a satisfactory explanation must consist of both counterfactual instances and a causal equation (or system of equations) that support the counterfactual instances. We will show that counterfactual instances by themselves explain little. We will further illustrate how explainable AI methods that provide both causal equations and counterfactual instances can successfully explain machine learning predictions.
翻译:在许多应用中,必须能够解释机器学习系统的决定。一种日益流行的方法是寻求提供\ emph{ 对应事实实例解释}。这些方法具体说明了与事实相反,一个人可以从机器学习系统得到其预期决定的近似世界。本文将借鉴科学哲学的文献,认为令人满意的解释必须既包括反事实事例,也包括支持反事实事例的因果等同(或方程式系统)。我们将表明反事实事例本身解释很少。我们将进一步说明提供因果等式和反事实事例的AI方法如何解释,能够成功地解释机器学习预测。