Explainable artificial intelligence techniques are evolving at breakneck speed, but suitable evaluation approaches currently lag behind. With explainers becoming increasingly complex and a lack of consensus on how to assess their utility, it is challenging to judge the benefit and effectiveness of different explanations. To address this gap, we take a step back from complex predictive algorithms and instead look into explainability of simple mathematical models. In this setting, we aim to assess how people perceive comprehensibility of different model representations such as mathematical formulation, graphical representation and textual summarisation (of varying scope). This allows diverse stakeholders -- engineers, researchers, consumers, regulators and the like -- to judge intelligibility of fundamental concepts that more complex artificial intelligence explanations are built from. This position paper charts our approach to establishing appropriate evaluation methodology as well as a conceptual and practical framework to facilitate setting up and executing relevant user studies.
翻译:可解释的人工智能技术正在以破碎的速度发展,但适当的评价方法目前落后。由于解释者越来越复杂,而且对如何评估其效用缺乏共识,因此判断不同解释的好处和有效性是困难的。为了弥补这一差距,我们从复杂的预测算法中退一步,而去研究简单数学模型的可解释性。在这一背景下,我们的目标是评估人们如何看待不同模型的可理解性,例如数学公式、图形代表制和(不同范围的)文字总结。这让不同的利益攸关方 -- -- 工程师、研究人员、消费者、监管者等 -- -- 能够判断更复杂的人工智能解释所依据的基本概念的不可理解性。本立场文件描绘了我们制定适当评价方法的方法,以及便利建立和实施相关用户研究的概念和实践框架。</s>