Stochastic gradient Markov chain Monte Carlo (SGMCMC) is a popular class of algorithms for scalable Bayesian inference. However, these algorithms include hyperparameters such as step size or batch size that influence the accuracy of estimators based on the obtained posterior samples. As a result, these hyperparameters must be tuned by the practitioner and currently no principled and automated way to tune them exists. Standard MCMC tuning methods based on acceptance rates cannot be used for SGMCMC, thus requiring alternative tools and diagnostics. We propose a novel bandit-based algorithm that tunes the SGMCMC hyperparameters by minimizing the Stein discrepancy between the true posterior and its Monte Carlo approximation. We provide theoretical results supporting this approach and assess various Stein-based discrepancies. We support our results with experiments on both simulated and real datasets, and find that this method is practical for a wide range of applications.
翻译:Stochatic 梯度 Markov 链 Monte Carlo(SGMC ) 是一种流行的可缩放贝叶色推算算法,但是,这些算法包括超参数,例如根据获得的远地点样本影响估计器准确性的级尺大小或批量大小,因此,这些超参数必须由执业者加以调整,目前不存在调和这些参数的有原则的自动方法。基于接受率的标准MC调控方法不能用于SGMCMC,因此需要替代工具和诊断。我们建议一种基于新颖的土匪算法,通过尽量减少真实的远地点与蒙特卡洛近距离之间的斯坦因差异来调控SGMCMC的双参数。我们提供理论结果支持这一方法,并评估各种基于斯坦基的差异。我们用模拟和真实数据集的实验结果来支持我们的结果,并发现这一方法对于广泛的应用是实用的。