In aspect-level sentiment classification (ASC), it is prevalent to equip dominant neural models with attention mechanisms, for the sake of acquiring the importance of each context word on the given aspect. However, such a mechanism tends to excessively focus on a few frequent words with sentiment polarities, while ignoring infrequent ones. In this paper, we propose a progressive self-supervised attention learning approach for neural ASC models, which automatically mines useful attention supervision information from a training corpus to refine attention mechanisms. Specifically, we iteratively conduct sentiment predictions on all training instances. Particularly, at each iteration, the context word with the maximum attention weight is extracted as the one with active/misleading influence on the correct/incorrect prediction of every instance, and then the word itself is masked for subsequent iterations. Finally, we augment the conventional training objective with a regularization term, which enables ASC models to continue equally focusing on the extracted active context words while decreasing weights of those misleading ones. Experimental results on multiple datasets show that our proposed approach yields better attention mechanisms, leading to substantial improvements over the two state-of-the-art neural ASC models. Source code and trained models are available at https://github.com/DeepLearnXMU/PSSAttention.
翻译:在心理层面的情绪分类(ASC)中,为获得对特定方面的每个上下文词的重要性,通常要为占主导地位的神经模型配备关注机制;然而,这种机制往往过分侧重于情绪极化的少许频繁的词,而忽视不常见的词;在本文中,我们提议对神经 ASC 模型采取渐进式自我监督的注意力学习方法,这些模型自动地利用培训资料中有用的注意力监督信息,以完善关注机制。具体地说,我们反复对所有培训实例进行感知预测。特别是,在每次迭代中,最受重视的字词被提取为对每个实例的正确/错误预测具有积极/误导影响,然后用隐蔽的词本身加以遮掩。最后,我们用一个正规化的术语来扩大常规培训目标,使ASC 模型能够同样地继续关注提取的积极字,同时降低这些误导词的重量。多个数据集的实验结果显示,我们所提议的方法能够产生更好的关注机制,导致两个州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州