Deep neural networks (DNNs) have become a proven and indispensable machine learning tool. As a black-box model, it remains difficult to diagnose what aspects of the model's input drive the decisions of a DNN. In countless real-world domains, from legislation and law enforcement to healthcare, such diagnosis is essential to ensure that DNN decisions are driven by aspects appropriate in the context of its use. The development of methods and studies enabling the explanation of a DNN's decisions has thus blossomed into an active, broad area of research. A practitioner wanting to study explainable deep learning may be intimidated by the plethora of orthogonal directions the field has taken. This complexity is further exacerbated by competing definitions of what it means ``to explain'' the actions of a DNN and to evaluate an approach's ``ability to explain''. This article offers a field guide to explore the space of explainable deep learning aimed at those uninitiated in the field. The field guide: i) Introduces three simple dimensions defining the space of foundational methods that contribute to explainable deep learning, ii) discusses the evaluations for model explanations, iii) places explainability in the context of other related deep learning research areas, and iv) finally elaborates on user-oriented explanation designing and potential future directions on explainable deep learning. We hope the guide is used as an easy-to-digest starting point for those just embarking on research in this field.
翻译:深心神经网络(DNNs)已经成为一个得到证明和不可或缺的机器学习工具。作为一个黑盒模型,仍然难以判断模型投入的哪些方面驱动DNN的决定。在无数现实世界领域,从立法和执法到医疗保健,这种诊断对于确保DNN决定的驱动由在其使用范围内适当的方面驱动至关重要。因此,有助于解释DNN决定的方法和研究的开发已经发展成一个活跃的、广泛的研究领域。作为黑盒模型的模型模型模型,想要学习深思熟虑的深思熟虑的实践者可能受到该字段所走的多处方向的恐吓。由于“解释DNN的行动”和“评价是否适合解释”这个方法的含义相互竞争的定义而进一步加剧了这种复杂性。这篇文章为探索以实地未启动者为目的的深思熟虑的深思熟虑空间提供了实地指南。 实地指南提出了三个简单的方面,界定基础方法的空间,有助于进行深入深入的深入学习,二)讨论关于开始的模型解释的评估,我们最终在研究中解释了这些研究领域采用的潜在方向。