Explainable AI is an emerging field providing solutions for acquiring insights into automated systems' rationale. It has been put on the AI map by suggesting ways to tackle key ethical and societal issues. Existing explanation techniques are often not comprehensible to the end user. Lack of evaluation and selection criteria also makes it difficult for the end user to choose the most suitable technique. In this study, we combine logic-based argumentation with Interpretable Machine Learning, introducing a preliminary meta-explanation methodology that identifies the truthful parts of feature importance oriented interpretations. This approach, in addition to being used as a meta-explanation technique, can be used as an evaluation or selection tool for multiple feature importance techniques. Experimentation strongly indicates that an ensemble of multiple interpretation techniques yields considerably more truthful explanations.
翻译:AI是一个新出现的领域,它为了解自动化系统的理由提供了解决办法,它通过提出解决关键伦理和社会问题的方法而被列入AI地图。现有的解释技术往往无法为终端用户所理解。缺乏评价和选择标准也使得终端用户难以选择最合适的技术。在这个研究中,我们把基于逻辑的论证与可解释的机器学习结合起来,引入了初步的元分析方法,确定了以特征为导向的解释的真实部分。这一方法除了用作元分析技术外,还可以用作多重特征重要技术的评价或选择工具。实验有力地表明,多种解释技术的组合可以产生更真实的解释。