The driving force behind the recent success of LSTMs has been their ability to learn complex and non-linear relationships. Consequently, our inability to describe these relationships has led to LSTMs being characterized as black boxes. To this end, we introduce contextual decomposition (CD), an interpretation algorithm for analysing individual predictions made by standard LSTMs, without any changes to the underlying model. By decomposing the output of a LSTM, CD captures the contributions of combinations of words or variables to the final prediction of an LSTM. On the task of sentiment analysis with the Yelp and SST data sets, we show that CD is able to reliably identify words and phrases of contrasting sentiment, and how they are combined to yield the LSTM's final prediction. Using the phrase-level labels in SST, we also demonstrate that CD is able to successfully extract positive and negative negations from an LSTM, something which has not previously been done.
翻译:LSTMs最近成功背后的驱动力是它们学习复杂和非线性关系的能力。 因此,我们无法描述这些关系导致LSTMs被定性为黑盒。 为此,我们引入了背景分解(CD),这是用来分析标准LSTMs所作的个别预测的一种解释算法,没有改变基本模型。通过分解LSTM的输出,CD捕捉了单词或变量组合对LSTM最后预测的贡献。关于与Yelp和SST数据集的情绪分析任务,我们显示CD能够可靠地识别反差情绪的文字和短语,以及它们如何结合得出LSTM的最后预测。我们使用SST中的语级标签,还表明CD能够成功地从LSTM中获取正反效果,这是以前没有做到的事情。