Topic models are one of the most frequently used models in machine learning due to its high interpretability and modular structure. However extending the model to include supervisory signal, incorporate pre-trained word embedding vectors and add nonlinear output function to the model is not an easy task because one has to resort to highly intricate approximate inference procedure. In this paper, we show that topic models could be viewed as performing a neighborhood aggregation algorithm where the messages are passed through a network defined over words. Under the network view of topic models, nodes corresponds to words in a document and edges correspond to either a relationship describing co-occurring words in a document or a relationship describing same word in the corpus. The network view allows us to extend the model to include supervisory signals, incorporate pre-trained word embedding vectors and add nonlinear output function to the model in a simple manner. Moreover, we describe a simple way to train the model that is well suited in a semi-supervised setting where we only have supervisory signals for some portion of the corpus and the goal is to improve prediction performance in the held-out data. Through careful experiments we show that our approach outperforms state-of-the-art supervised Latent Dirichlet Allocation implementation in both held-out document classification tasks and topic coherence.
翻译:主题模型是机器学习中最常用的模型之一,因为其可解释性和模块结构较高。但将模型扩展至包括监督信号,纳入预先训练的文字嵌入矢量,并将非线性输出功能添加到模型中,这不是一件容易的任务,因为人们不得不采用高度复杂的近似推导程序。在本文中,我们表明,主题模型可以被视为是执行邻里汇总算法,通过以文字定义的网络网络传递信息。在对主题模型的网络观察中,节点与文档中的单词相对应,边缘对应于描述文档中共同出现单词或描述本体中相同单词的关系。网络视图允许我们扩展模型,以包括监督信号,纳入预先训练的字嵌入矢量,并以简单的方式将非线性输出函数添加到模型中。此外,我们描述了一种简单的方法来培训模型,该模型在半超强的环境下非常适合使用该元素的某些部分的监管信号,目标是改进被搁置数据的预测性工作。通过仔细的实验,我们展示了我们的方法超越了Dirmat-Streform lad laft laft laft laft laft laft laft laft laction laction laft laction laction laction laction laction laction laft laft laft laft laction laft laft lax