The performance of modern algorithms on certain computer vision tasks such as object recognition is now close to that of humans. This success was achieved at the price of complicated architectures depending on millions of parameters and it has become quite challenging to understand how particular predictions are made. Interpretability methods propose to give us this understanding. In this paper, we study LIME, perhaps one of the most popular. On the theoretical side, we show that when the number of generated examples is large, LIME explanations are concentrated around a limit explanation for which we give an explicit expression. We further this study for elementary shape detectors and linear models. As a consequence of this analysis, we uncover a connection between LIME and integrated gradients, another explanation method. More precisely, the LIME explanations are similar to the sum of integrated gradients over the superpixels used in the preprocessing step of LIME.
翻译:某些计算机视觉任务(如物体识别)的现代算法的性能现在接近于人类。这种成功是以复杂结构的价格取得的,其价格取决于数百万参数,了解具体预测是如何作出的,已经变得相当困难。解释方法建议给我们这种理解。在本文中,我们研究LIME,也许是最受欢迎的一个。在理论方面,我们表明,当生成的例子数量巨大时,LIME的解释集中在一个我们明确表达的有限解释上。我们进一步研究基本形状探测器和线性模型。由于这一分析,我们发现了LIME与集成梯度之间的联系,这是另一种解释方法。更确切地说,LIME的解释类似于LIME预处理步骤中使用的超像素的综合梯数。