With the broader and highly successful usage of machine learning in industry and the sciences, there has been a growing demand for Explainable AI. Interpretability and explanation methods for gaining a better understanding about the problem solving abilities and strategies of nonlinear Machine Learning, in particular, deep neural networks, are therefore receiving increased attention. In this work we aim to (1) provide a timely overview of this active emerging field, with a focus on 'post-hoc' explanations, and explain its theoretical foundations, (2) put interpretability algorithms to a test both from a theory and comparative evaluation perspective using extensive simulations, (3) outline best practice aspects i.e. how to best include interpretation methods into the standard usage of machine learning and (4) demonstrate successful usage of explainable AI in a representative selection of application scenarios. Finally, we discuss challenges and possible future directions of this exciting foundational field of machine learning.
翻译:随着在工业和科学中更广泛和非常成功地使用机器学习,对可解释的AI的需求日益增长。因此,人们日益重视为更好地了解非线性机器学习能力和战略,特别是深神经网络的问题而采用的解释和解释方法。在这项工作中,我们的目标是:(1) 及时概述这个活跃的新兴领域,重点是“热后”解释,并解释其理论基础;(2) 从理论和比较评价角度,利用广泛的模拟,将可解释性算法置于测试之中;(3) 概述最佳做法方面,即如何最好地将解释方法纳入机器学习的标准使用;(4) 展示在有代表性地选择应用情景时成功使用可解释的AI。最后,我们讨论了这一令人振奋的机器学习基础领域的挑战和未来可能的方向。