In aspect-based sentiment analysis (ABSA), many neural models are equipped with an attention mechanism to quantify the contribution of each context word to sentiment prediction. However, such a mechanism suffers from one drawback: only a few frequent words with sentiment polarities are tended to be taken into consideration for final sentiment decision while abundant infrequent sentiment words are ignored by models. To deal with this issue, we propose a progressive self-supervised attention learning approach for attentional ABSA models. In this approach, we iteratively perform sentiment prediction on all training instances, and continually learn useful attention supervision information in the meantime. During training, at each iteration, context words with the highest impact on sentiment prediction, identified based on their attention weights or gradients, are extracted as words with active/misleading influence on the correct/incorrect prediction for each instance. Words extracted in this way are masked for subsequent iterations. To exploit these extracted words for refining ABSA models, we augment the conventional training objective with a regularization term that encourages ABSA models to not only take full advantage of the extracted active context words but also decrease the weights of those misleading words. We integrate the proposed approach into three state-of-the-art neural ABSA models. Experiment results and in-depth analyses show that our approach yields better attention results and significantly enhances the performance of all three models. We release the source code and trained models at https://github.com/DeepLearnXMU/PSSAttention.
翻译:在基于情感的方面分析(ABSA)中,许多神经模型都配备了一种关注机制,以量化每个背景词对情绪预测的贡献,然而,这种机制有一个缺陷:在最后的情感决定中,只考虑情绪极化的少许频繁的单词,而模型则忽视大量不常见的情绪单词。为了处理这一问题,我们建议对ABSA的焦点模型采取渐进式自我监督的注意力学习方法。在这个方法中,我们反复对所有培训实例进行情感预测,并同时不断学习有用的关注监督信息。在每次循环培训中,根据关注权重或梯度确定的对情绪预测影响最大的背景单词,作为对每种情况的正确/错误预测的积极/误导影响词被采纳。我们用这种方式提取的单词掩盖了以后的反复。为了利用这些提取的单词来改进ABSA的模型,我们用一个正规化的术语来扩大常规培训目标,鼓励ABSA的模型不仅充分利用所提取的积极上的背景词,而且还减少基于情绪预测力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力判断力分析力分析力判断力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力判断力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力分析力