Though with progress, model learning and performing posterior inference still remains a common challenge for using deep generative models, especially for handling discrete hidden variables. This paper is mainly concerned with algorithms for learning Helmholz machines, which is characterized by pairing the generative model with an auxiliary inference model. A common drawback of previous learning algorithms is that they indirectly optimize some bounds of the targeted marginal log-likelihood. In contrast, we successfully develop a new class of algorithms, based on stochastic approximation (SA) theory of the Robbins-Monro type, to directly optimize the marginal log-likelihood and simultaneously minimize the inclusive KL-divergence. The resulting learning algorithm is thus called joint SA (JSA). Moreover, we construct an effective MCMC operator for JSA. Our results on the MNIST datasets demonstrate that the JSA's performance is consistently superior to that of competing algorithms like RWS, for learning a range of difficult models.
翻译:尽管取得了进步,但模型学习和进行事后推断仍然是使用深层基因模型的一个常见挑战,特别是用于处理离散隐藏变量。本文件主要涉及学习Helmholz机器的算法,其特点是将基因模型与辅助推论模型配对。以往学习算法的一个常见缺点是,它们间接优化了目标边际日志相似性的某些界限。相比之下,我们成功地根据Robbins-Monro类型的随机近似(SA)理论开发了一种新的算法,以直接优化边际日志相似性,同时尽量减少包容性的KL-diverence。因此,由此产生的学习算法被称为联合SA(JSA )。此外,我们为JSA(JSA ) 构建了一个有效的MC MMC操作员。我们在MIT数据集上的结果表明,JSA的性能始终优于RWS等相互竞争的算法,以学习一系列困难的模型。