Gaussian processes (GPs) are frequently used in machine learning and statistics to construct powerful models. However, when employing GPs in practice, important considerations must be made, regarding the high computational burden, approximation of the posterior, choice of the covariance function and inference of its hyperparmeters. To address these issues, Hensman et al. (2015) combine variationally sparse GPs with Markov chain Monte Carlo (MCMC) to derive a scalable, flexible and general framework for GP models. Nevertheless, the resulting approach requires intractable likelihood evaluations for many observation models. To bypass this problem, we propose a pseudo-marginal (PM) scheme that offers asymptotically exact inference as well as computational gains through doubly stochastic estimators for the intractable likelihood and large datasets. In complex models, the advantages of the PM scheme are particularly evident, and we demonstrate this on a two-level GP regression model with a nonparametric covariance function to capture non-stationarity.
翻译:Gausian进程(GPs)经常用于机器学习和统计,以构建强大的模型;然而,在实际使用GPs时,必须就高计算负担、近似后台、选择共变函数及其超光度计的推论等考虑重要;为解决这些问题,Hensman等人(2015年)将分散的GPs与Markov链Monte Carlo(MCMC)相结合,为GP模式制定可伸缩、灵活和一般的框架;然而,由此形成的方法要求对许多观测模型进行棘手的概率评估。为了绕过这一问题,我们提议了一个假边际(PM)方案,通过对难用的可能性和大型数据集进行双分精确的估算,提供随机精确的推论以及计算收益。在复杂的模型中,PM计划的优势特别明显,我们用两种具有非对称共变函数的GP回归模型来显示非常态性。