This work focuses on combining nonparametric topic models with Auto-Encoding Variational Bayes (AEVB). Specifically, we first propose iTM-VAE, where the topics are treated as trainable parameters and the document-specific topic proportions are obtained by a stick-breaking construction. The inference of iTM-VAE is modeled by neural networks such that it can be computed in a simple feed-forward manner. We also describe how to introduce a hyper-prior into iTM-VAE so as to model the uncertainty of the prior parameter. Actually, the hyper-prior technique is quite general and we show that it can be applied to other AEVB based models to alleviate the {\it collapse-to-prior} problem elegantly. Moreover, we also propose HiTM-VAE, where the document-specific topic distributions are generated in a hierarchical manner. HiTM-VAE is even more flexible and can generate topic distributions with better variability. Experimental results on 20News and Reuters RCV1-V2 datasets show that the proposed models outperform the state-of-the-art baselines significantly. The advantages of the hyper-prior technique and the hierarchical model construction are also confirmed by experiments.
翻译:这项工作侧重于将非参数性专题模型与自动编码变异贝壳(AEVB)相结合。 具体地说,我们首先建议iTM-VAE, 将这些专题作为可培训的参数处理,而文件特定专题比例则通过粉碎的构造获得。iTM-VAE的推论是由神经网络模拟的,这样就可以以简单的进取-向方式计算出。我们还描述了如何在iTM-VAE中引入超主要专题模型,以模拟先前参数的不确定性。事实上,超主要技术相当笼统,我们表明它可以适用于其他基于AEVB的模型,以优雅地缓解“倒塌-主要”问题。此外,我们还提议HTM-VAE, 以等级方式生成文件特定专题分布。 HTM-VAE甚至更加灵活,能够产生更易变性的专题分布。 20News和路透社RCV1-V2号的实验结果显示,拟议的模型也大大超越了高等级性实验性基准。