Explainable AI was born as a pathway to allow humans to explore and understand the inner working of complex systems. But establishing what is an explanation and objectively evaluating explainability, are not trivial tasks. With this paper, we present a new model-agnostic metric to measure the Degree of Explainability of information in an objective way, exploiting a specific theoretical model from Ordinary Language Philosophy called the Achinstein's Theory of Explanations, implemented with an algorithm relying on deep language models for knowledge graph extraction and information retrieval. In order to understand whether this metric is actually behaving as explainability is expected to, we devised a few experiments and user-studies involving more than 160 participants, evaluating two realistic AI-based systems for healthcare and finance using famous AI technology including Artificial Neural Networks and TreeSHAP. The results we obtained are statistically significant (with P values lower than 0.01), suggesting that our proposed metric for measuring the Degree of Explainability is robust on several scenarios and it aligns with concrete expectations.
翻译:可以解释的大赦国际是人类探索和理解复杂系统内部工作的一种途径。但是,确定什么是解释,客观地评估解释性,并不是一件微不足道的任务。我们用这份文件提出了一个新的模型-不可知性衡量标准,以客观的方式衡量信息的可解释性程度,利用普通语言哲学中称为“阿钦斯坦解释理论”的具体理论模型,该模型采用一种基于深语言模型的算法,用于知识图提取和信息检索。为了了解这一计量是否实际上可以解释,我们设计了一些实验和用户研究,有160多名参与者参与,用著名的AI技术,包括人工神经网络和TreaSHAP,评估了两种现实的基于AI的医疗保健和融资系统。我们获得的结果具有统计意义(P值低于0.01),表明我们提议的用于衡量可解释性程度的计量标准在若干情景上是稳健的,符合具体期望。