Layer-wise relevance propagation (LRP) is a widely used and powerful technique to reveal insights into various artificial neural network (ANN) architectures. LRP is often used in the context of image classification. The aim is to understand, which parts of the input sample have highest relevance and hence most influence on the model prediction. Relevance can be traced back through the network to attribute a certain score to each input pixel. Relevance scores are then combined and displayed as heat maps and give humans an intuitive visual understanding of classification models. Opening the black box to understand the classification engine in great detail is essential for domain experts to gain trust in ANN models. However, there are pitfalls in terms of model-inherent artifacts included in the obtained relevance maps, that can easily be missed. But for a valid interpretation, these artifacts must not be ignored. Here, we apply and revise LRP on various ANN architectures trained as classifiers on geospatial and synthetic data. Depending on the network architecture, we show techniques to control model focus and give guidance to improve the quality of obtained relevance maps to separate facts from artifacts.
翻译:地层关联性传播(LRP)是一种广泛使用和强大的技术,可以揭示各种人工神经网络结构的洞察力。LRP经常用于图像分类。目的是了解输入样本的哪些部分具有最高的相关性,从而对模型预测产生最大影响。可以通过网络追溯相关性,将某个分数归属于每个输入像素。相关性分数随后作为热图加以合并并显示,使人类对分类模型有一个直观的直观理解。打开黑盒以深入了解分类引擎对于域专家信任ANN模型至关重要。然而,在获得的关联性地图中包含的模型固有文物方面有一些陷阱,很容易被忽略。但是,对于有效的解释,这些文物绝不能被忽略。在这里,我们应用并修改了作为地理空间和合成数据分类师培训的各种ANN结构。根据网络结构,我们展示了控制模型焦点的技术,并提供指导,以提高获得的关联性地图的质量,使之与手工艺品分开。</s>