A variety of methods have been proposed to try to explain how deep neural networks make their decisions. Key to those approaches is the need to sample the pixel space efficiently in order to derive importance maps. However, it has been shown that the sampling methods used to date introduce biases and other artifacts, leading to inaccurate estimates of the importance of individual pixels and severely limit the reliability of current explainability methods. Unfortunately, the alternative -- to exhaustively sample the image space is computationally prohibitive. In this paper, we introduce EVA (Explaining using Verified perturbation Analysis) -- the first explainability method guarantee to have an exhaustive exploration of a perturbation space. Specifically, we leverage the beneficial properties of verified perturbation analysis -- time efficiency, tractability and guaranteed complete coverage of a manifold -- to efficiently characterize the input variables that are most likely to drive the model decision. We evaluate the approach systematically and demonstrate state-of-the-art results on multiple benchmarks.
翻译:提出了多种方法,试图解释深神经网络是如何作出决定的。这些方法的关键是需要有效地对像素空间进行抽样,以便绘制重要地图。然而,已经表明,迄今为止采用的抽样方法引入了偏差和其他文物,导致对单个像素重要性的不准确估计,严重限制了当前解释方法的可靠性。不幸的是,对图像空间进行详尽抽样的替代办法是计算上令人窒息的。在本文件中,我们引入了EVA(利用经过验证的扰动分析来解释) -- -- 这是保证对扰动空间进行彻底探索的第一个解释性方法。具体地说,我们利用经过核实的扰动分析的有益特性 -- -- 时间效率、可变性和保证全面覆盖一个方位 -- -- 来有效地描述最有可能推动示范决定的投入变量。我们系统地评估该方法,并展示在多个基准方面的最新结果。