Recurrent neural networks are a standard building block in numerous machine learning domains, from natural language processing to time-series classification. While their application has grown ubiquitous, understanding of their inner workings is still lacking. In practice, the complex decision-making in these models is seen as a black-box, creating a tension between accuracy and interpretability. Moreover, the ability to understand the reasoning process of a model is important in order to debug it and, even more so, to build trust in its decisions. Although considerable research effort has been guided towards explaining black-box models in recent years, recurrent models have received relatively little attention. Any method that aims to explain decisions from a sequence of instances should assess, not only feature importance, but also event importance, an ability that is missing from state-of-the-art explainers. In this work, we contribute to filling these gaps by presenting TimeSHAP, a model-agnostic recurrent explainer that leverages KernelSHAP's sound theoretical footing and strong empirical results. As the input sequence may be arbitrarily long, we further propose a pruning method that is shown to dramatically improve its efficiency in practice.
翻译:经常性神经网络是许多机器学习领域的标准基石,从自然语言处理到时间序列分类,从自然语言处理到时间序列分类,其应用已变得无处不在,但对其内在功能的理解仍然缺乏。实际上,这些模型的复杂决策被视为黑箱,在准确性和可解释性之间造成紧张。此外,理解模型推理过程的能力对于调试模型十分重要,甚至对于建立对其决定的信任也很重要。尽管近年来大量研究努力一直引导着解释黑盒模型,但经常模式却很少受到注意。任何旨在解释一系列案例决定的方法,不仅应该评估其重要性,而且应该评估其重要性,并且应当评估其重要性,从最新解释器中缺少的能力。在这项工作中,我们通过展示一个模型-认知性经常性解释器来填补这些差距,利用KernelSHAP的可靠理论基础和强有力的实证结果。由于输入序列可能是任意性的,我们进一步提议一种支线方法,以大幅提高实践效率。