We propose a Topic Compositional Neural Language Model (TCNLM), a novel method designed to simultaneously capture both the global semantic meaning and the local word ordering structure in a document. The TCNLM learns the global semantic coherence of a document via a neural topic model, and the probability of each learned latent topic is further used to build a Mixture-of-Experts (MoE) language model, where each expert (corresponding to one topic) is a recurrent neural network (RNN) that accounts for learning the local structure of a word sequence. In order to train the MoE model efficiently, a matrix factorization method is applied, by extending each weight matrix of the RNN to be an ensemble of topic-dependent weight matrices. The degree to which each member of the ensemble is used is tied to the document-dependent probability of the corresponding topics. Experimental results on several corpora show that the proposed approach outperforms both a pure RNN-based model and other topic-guided language models. Further, our model yields sensible topics, and also has the capacity to generate meaningful sentences conditioned on given topics.
翻译:我们提议了一个专题构成神经语言模型(TCNLM),这是一个新颖的方法,旨在同时捕捉全球语义含义和文件中本地词顺序结构。TCNLM通过神经主题模型学习文件的全球语义一致性,每个已学过的潜在主题的概率被进一步用于构建一个混合专家语言模型(MOE),在这个模型中,每个专家(对应一个专题)是一个经常性神经网络(RNNN),该神经网络核算了一个单词序列的本地结构。为了高效地培训教育部模型,采用了矩阵因子化方法,通过扩展RNN的每个重量矩阵,成为依赖主题的加权矩阵的共集体。使用该元素每个成员的程度与相应专题的文件依赖概率挂钩。几个公司实验结果显示,拟议的方法超越了纯RNN模式和其他主题指导语言模型。此外,我们模型生成了明智的专题,并具备了生成有意义条件判词的能力。