We present a novel method for reliably explaining the predictions of neural networks. We consider an explanation reliable if it identifies input features relevant to the model output by considering the input and the neighboring data points. Our method is built on top of the assumption of smooth landscape in a loss function of the model prediction: locally consistent loss and gradient profile. A theoretical analysis established in this study suggests that those locally smooth model explanations are learned using a batch of noisy copies of the input with the L1 regularization for a saliency map. Extensive experiments support the analysis results, revealing that the proposed saliency maps retrieve the original classes of adversarial examples crafted against both naturally and adversarially trained models, significantly outperforming previous methods. We further demonstrated that such good performance results from the learning capability of this method to identify input features that are truly relevant to the model output of the input and the neighboring data points, fulfilling the requirements of a reliable explanation.
翻译:我们提出了一个可靠解释神经网络预测的新方法。我们认为,如果通过考虑输入和相邻数据点来确定与模型输出相关的输入特征,解释是可靠的。我们的方法建立在模型预测损失函数中光滑地貌假设之上:地方一致的损失和梯度剖面。本研究中建立的一项理论分析表明,这些本地平稳的模型解释是利用一组与突出的地图L1正规化相关的输入的杂音拷贝来学习的。广泛的实验支持了分析结果,表明拟议的突出地图检索了针对自然和对抗性训练的模型所制作的最初类别的对抗性范例,大大超过以往的方法。我们进一步表明,这一方法的学习能力取得了良好的绩效,从而确定了与输入和相邻数据点的模型输出真正相关的投入特征,满足了可靠解释的要求。