Explainable artificial intelligence techniques are developed at breakneck speed, but suitable evaluation approaches lag behind. With explainers becoming increasingly complex and a lack of consensus on how to assess their utility, it is challenging to judge the benefit and effectiveness of different explanations. To address this gap, we take a step back from sophisticated predictive algorithms and instead look into explainability of simple decision-making models. In this setting, we aim to assess how people perceive comprehensibility of their different representations such as mathematical formulation, graphical representation and textual summarisation (of varying complexity and scope). This allows us to capture how diverse stakeholders -- engineers, researchers, consumers, regulators and the like -- judge intelligibility of fundamental concepts that more elaborate artificial intelligence explanations are built from. This position paper charts our approach to establishing appropriate evaluation methodology as well as a conceptual and practical framework to facilitate setting up and executing relevant user studies.
翻译:可解释的人工智能技术正以惊人的速度发展,但适当的评估方法却落后于此。随着解释器越来越复杂,评估其效用缺乏共识,评估不同解释的利益和有效性就变得困难。为了填补这一空白,我们从复杂的预测算法中抽身而出,而是研究简单决策模型的可解释性。在这个设置中,我们的目标是评估人们如何知觉不同表示的可理解性,例如数学公式、图形表示和文本摘要(复杂性和范围不同)。这使我们能够捕捉到不同利益相关者——工程师、研究人员、消费者、监管机构等——对更为详尽的人工智能解释所构建的基本概念的智能判断。本文阐述了我们设立适当的评估方法以及构建概念和实践框架以促进设置和执行相关用户研究的方法。