Explainable AI was born as a pathway to allow humans to explore and understand the inner working of complex systems. But establishing what is an explanation and objectively evaluating explainability, are not trivial tasks. With this paper, we present a new model-agnostic metric to measure the Degree of Explainability of information in an objective way, exploiting a specific theoretical model from Ordinary Language Philosophy called the Achinstein's Theory of Explanations, implemented with an algorithm relying on deep language models for knowledge graph extraction and information retrieval. To understand whether this metric is actually able to measure explainability, we devised a few experiments and user studies involving more than 190 participants, evaluating two realistic systems for healthcare and finance using famous AI technology including Artificial Neural Networks and TreeSHAP. The results we obtained are statistically significant (with p-values lower than .01), suggesting that our proposed metric for measuring the Degree of Explainability is robust on several scenarios and it aligns with concrete expectations.
翻译:可以解释的大赦国际是人类探索和理解复杂系统内部工作的一种途径。但是,确定什么是解释,客观地评估解释性,并不是一件微不足道的任务。我们用这份文件提出了一个新的模型-不可知性衡量标准,以客观的方式衡量信息的可解释性程度,利用普通语言哲学中称为“阿钦斯坦解释理论”的具体理论模型,该模型采用一种依赖深语言模型的算法,用于知识图提取和信息检索。为了了解这一指标是否真正能够测量可解释性,我们设计了几项实验和用户研究,有190多名参与者参与,利用著名的人工神经网络和树本健康评估了两种现实的医疗保健和融资系统。我们获得的结果具有统计意义(其价值小于.01),表明我们提议的用于衡量可解释性程度的参数在几种情景上是稳健的,符合具体期望。