Topic models have been widely explored as probabilistic generative models of documents. Traditional inference methods have sought closed-form derivations for updating the models, however as the expressiveness of these models grows, so does the difficulty of performing fast and accurate inference over their parameters. This paper presents alternative neural approaches to topic modelling by providing parameterisable distributions over topics which permit training by backpropagation in the framework of neural variational inference. In addition, with the help of a stick-breaking construction, we propose a recurrent network that is able to discover a notionally unbounded number of topics, analogous to Bayesian non-parametric topic models. Experimental results on the MXM Song Lyrics, 20NewsGroups and Reuters News datasets demonstrate the effectiveness and efficiency of these neural topic models.
翻译:传统推论方法为更新模型寻求了封闭式推论,然而,随着这些模型的表达性增加,对这些模型的参数进行快速和准确推论的困难也随之增加。本文介绍了对专题建模的替代神经方法,为在神经变异推理框架内进行反光反光分析培训的专题提供可参数分布。此外,在粘土构造的帮助下,我们提议建立一个经常性网络,能够发现一些概念上不受限制的专题,类似于巴伊西亚非参数性专题模型。MXM Song Lyrics、20News Groups和路透社新闻数据集的实验结果显示了这些神经专题模型的有效性和效率。