A longstanding challenge surrounding deep learning algorithms is unpacking and understanding how they make their decisions. Explainable Artificial Intelligence (XAI) offers methods to provide explanations of internal functions of algorithms and reasons behind their decisions in ways that are interpretable and understandable to human users. . Numerous XAI approaches have been developed thus far, and a comparative analysis of these strategies seems necessary to discern their relevance to clinical prediction models. To this end, we first implemented two prediction models for short- and long-term outcomes of traumatic brain injury (TBI) utilizing structured tabular as well as time-series physiologic data, respectively. Six different interpretation techniques were used to describe both prediction models at the local and global levels. We then performed a critical analysis of merits and drawbacks of each strategy, highlighting the implications for researchers who are interested in applying these methodologies. The implemented methods were compared to one another in terms of several XAI characteristics such as understandability, fidelity, and stability. Our findings show that SHAP is the most stable with the highest fidelity but falls short of understandability. Anchors, on the other hand, is the most understandable approach, but it is only applicable to tabular data and not time series data.
翻译:围绕深层次学习算法的长期挑战是解开和理解它们如何作出决定。可以解释的人工智能(XAI)提供了各种方法来解释算法的内部功能和其决定背后的理由,其方式可以解释和理解人类用户。迄今为止,已经制定了许多XAI方法,对这些战略进行比较分析似乎是必要的,以辨别它们与临床预测模型的相关性。为此,我们首先使用结构化的表格和时间序列物理数据对创伤性脑损伤的短期和长期结果进行了两种预测模型。使用了六种不同的解释技术来描述地方和全球两级的预测模型。我们随后对每项战略的优点和缺点进行了批判性分析,突出了对有兴趣应用这些方法的研究人员的影响。所实施的方法在可理解性、忠诚性和稳定性等几个XAI特征方面被比较为另一个特征。我们的研究结果表明,SHAP是最高忠诚度最稳定但无法理解性最差的。另一端,紧要者是最难理解性的方法,但数据只适用于列表。