Stochastic gradient Markov chain Monte Carlo (SGMCMC) has become a popular method for scalable Bayesian inference. These methods are based on sampling a discrete-time approximation to a continuous time process, such as the Langevin diffusion. When applied to distributions defined on a constrained space, such as the simplex, the time-discretisation error can dominate when we are near the boundary of the space. We demonstrate that while current SGMCMC methods for the simplex perform well in certain cases, they struggle with sparse simplex spaces; when many of the components are close to zero. However, most popular large-scale applications of Bayesian inference on simplex spaces, such as network or topic models, are sparse. We argue that this poor performance is due to the biases of SGMCMC caused by the discretization error. To get around this, we propose the stochastic CIR process, which removes all discretization error and we prove that samples from the stochastic CIR process are asymptotically unbiased. Use of the stochastic CIR process within a SGMCMC algorithm is shown to give substantially better performance for a topic model and a Dirichlet process mixture model than existing SGMCMC approaches.
翻译:石化梯度 Markov 链 Monte Carlo (SGMC ) 已成为一种流行的可缩放贝叶斯人的推算方法。 这些方法基于对离散时间近似到连续时间过程的取样, 如 Langevin 扩散。 当应用到限制空间定义的分布时, 如简单x, 当我们接近空间边界时, 时间分解错误会占主导地位 。 我们证明当前 SGMC 简单x 的SGMC 方法在某些情况下效果良好, 它们与稀疏的简单x 空间搏斗; 当许多部件接近于零时 。 然而, 在简单x 空间, 如网络或主题模型上, 最受欢迎的巴伊西亚人大规模推论应用非常少 。 我们争论说, 这种不良的性能是由于离散错误导致的 SGMC 偏差。 为了绕过这个过程, 我们建议采用SGMC 模型化 CIR 进程, 消除所有离散错误, 并且我们证明从 Schatic CIR 进程中提取的样品是不带偏见的。 在 SGMC MIC 的模型中使用一个比 模型模型模型模型的模型模型模型模型模型化进程, 显示SGMC 。