Text data are increasingly handled in an automated fashion by machine learning algorithms. But the models handling these data are not always well-understood due to their complexity and are more and more often referred to as "black-boxes." Interpretability methods aim to explain how these models operate. Among them, LIME has become one of the most popular in recent years. However, it comes without theoretical guarantees: even for simple models, we are not sure that LIME behaves accurately. In this paper, we provide a first theoretical analysis of LIME for text data. As a consequence of our theoretical findings, we show that LIME indeed provides meaningful explanations for simple models, namely decision trees and linear models.
翻译:文本数据越来越多地由机器学习算法自动处理。 但是,处理这些数据的模型由于复杂,并非总能很好地理解,而且越来越经常被称为“黑箱 ” 。 解释方法旨在解释这些模型是如何运作的。 其中,LIME近年来已成为最受欢迎的模式之一。 但是,它没有理论保证:即使对于简单的模型,我们也不确定LIME的行为是否准确。在本文中,我们对LIME的文本数据进行了初步的理论分析。由于我们的理论发现,我们证明LIME确实为简单的模型,即决定树和线性模型提供了有意义的解释。