Nowadays, it is growing interest to make Machine Learning (ML) systems more understandable and trusting to general users. Thus, generating explanations for ML system behaviours that are understandable to human beings is a central scientific and technological issue addressed by the rapidly growing research area of eXplainable Artificial Intelligence (XAI). Recently, it is becoming more and more evident that new directions to create better explanations should take into account what a good explanation is to a human user, and consequently, develop XAI solutions able to provide user-centred explanations. This paper suggests taking advantage of developing an XAI general approach that allows producing explanations for an ML system behaviour in terms of different and user-selected input features, i.e., explanations composed of input properties that the human user can select according to his background knowledge and goals. To this end, we propose an XAI general approach which is able: 1) to construct explanations in terms of input features that represent more salient and understandable input properties for a user, which we call here Middle-Level input Features (MLFs), 2) to be applied to different types of MLFs. We experimentally tested our approach on two different datasets and using three different types of MLFs. The results seem encouraging.
翻译:目前,人们越来越有兴趣使机器学习系统更易理解,更信任普通用户,因此,为ML系统行为提供对人类可以理解的解释,这是电子可移植人工智能(XAI)研究领域迅速增长,解决了一个核心的科学和技术问题。最近,越来越明显的一点是,为创造更好的解释而提出的新方向应该考虑到对人类用户来说什么是一个很好的解释,从而开发能够提供以用户为中心的解释的XAI解决方案。本文建议利用开发XAI通用方法,以便能够以不同和用户选择的投入特点,即由输入特性构成的解释,而人类用户可以根据其背景知识和目标选择输入特性。为此,我们提议XAI通用方法能够:(1) 以代表更突出和易懂的输入特性的输入特性来解释用户,我们在这里称之为中等级输入特性,(2) 用于不同类型的MLF。我们实验了两种不同的数据组合和三种不同的LF的结果。我们似乎用不同的LF类型来鼓励两种不同的LF。