In recent years, an abundance of feature attribution methods for explaining neural networks have been developed. Especially in the field of computer vision, many methods for generating saliency maps providing pixel attributions exist. However, their explanations often contradict each other and it is not clear which explanation to trust. A natural solution to this problem is the aggregation of multiple explanations. We present and compare different pixel-based aggregation schemes with the goal of generating a new explanation, whose fidelity to the model's decision is higher than each individual explanation. Using methods from the field of Bayesian Optimization, we incorporate the variance between the individual explanations into the aggregation process. Additionally, we analyze the effect of multiple normalization techniques on ensemble aggregation.
翻译:近年来,在解释神经网络方面发展了丰富的特征归属方法,特别是在计算机视觉领域,有许多生成提供像素属性特征的突出地图的方法,然而,这些解释往往相互矛盾,对信任的解释也不明确。这个问题的自然解决办法是多种解释的汇总。我们提出并比较不同的像素汇总计划,目的是产生新的解释,这种解释对模型决定的忠诚程度高于每一项个别解释。我们采用巴耶西亚最佳化领域的方法,将个别解释之间的差异纳入聚合过程。此外,我们分析了多重正常化技术对共性聚合的影响。