Unsupervised aspect detection (UAD) aims at automatically extracting interpretable aspects and identifying aspect-specific segments (such as sentences) from online reviews. However, recent deep learning-based topic models, specifically aspect-based autoencoder, suffer from several problems, such as extracting noisy aspects and poorly mapping aspects discovered by models to the aspects of interest. To tackle these challenges, in this paper, we first propose a self-supervised contrastive learning framework and an attention-based model equipped with a novel smooth self-attention (SSA) module for the UAD task in order to learn better representations for aspects and review segments. Secondly, we introduce a high-resolution selective mapping (HRSMap) method to efficiently assign aspects discovered by the model to aspects of interest. We also propose using a knowledge distilling technique to further improve the aspect detection performance. Our methods outperform several recent unsupervised and weakly supervised approaches on publicly available benchmark user review datasets. Aspect interpretation results show that extracted aspects are meaningful, have good coverage, and can be easily mapped to aspects of interest. Ablation studies and attention weight visualization also demonstrate the effectiveness of SSA and the knowledge distilling method.
翻译:未经监督的方面探测(UAD)旨在自动提取可解释的方面,并从网上审查中确定特定部分(如句子),然而,最近深深层次的学习主题模型,特别是基于侧面的自动编码器,遇到了若干问题,例如提取一些吵闹的方面,模型发现到感兴趣的方面,绘图不善。为了应对这些挑战,我们首先在本文件中提议一个自我监督的对比学习框架和基于关注的模式,为UAD任务配备一个全新的顺利自省模块,以学习对方面和审查部分的更好表述。第二,我们采用一种高分辨率的选择性绘图(HRSMAP)方法,以便有效地将模型发现的方面分配给感兴趣的方面。我们还提议使用一种知识蒸馏技术来进一步改进方面探测性能。我们的方法超越了最近在公开的基准用户审查数据集上一些未经监督和监督的方法。通过推断的结果表明,提取的方面是有意义的,覆盖面良好,并且可以很容易地与感兴趣的方面进行绘图。