Approximate Markov chain Monte Carlo (MCMC) offers the promise of more rapid sampling at the cost of more biased inference. Since standard MCMC diagnostics fail to detect these biases, researchers have developed computable Stein discrepancy measures that provably determine the convergence of a sample to its target distribution. This approach was recently combined with the theory of reproducing kernels to define a closed-form kernel Stein discrepancy (KSD) computable by summing kernel evaluations across pairs of sample points. We develop a theory of weak convergence for KSDs based on Stein's method, demonstrate that commonly used KSDs fail to detect non-convergence even for Gaussian targets, and show that kernels with slowly decaying tails provably determine convergence for a large class of target distributions. The resulting convergence-determining KSDs are suitable for comparing biased, exact, and deterministic sample sequences and simpler to compute and parallelize than alternative Stein discrepancies. We use our tools to compare biased samplers, select sampler hyperparameters, and improve upon existing KSD approaches to one-sample hypothesis testing and sample quality improvement.
翻译:由于标准的MCMC诊断未能检测出这些偏差,研究人员制定了可比较的 Stein 差异性测量方法,可以确定样本与目标分布的趋同程度。这一方法最近与再生产内核理论相结合,以确定封闭式内核的内核差异(KSD),通过对两个抽样点进行内核评价,可以对封闭式内核差异进行比较。我们根据Stein的方法为KSDs开发了一种衰弱的趋同理论,表明常用的KSDs甚至未能检测到高斯目标的非趋同性,并表明带有缓慢腐蚀尾巴的内核部分可以可确定大量目标分布的趋同性。由此产生的趋同-确定KSDs适合于比较偏差、精确和确定性样品序列,比替代性 Stein差异更便于比较和相平行。我们使用我们的工具来比较偏差的取样器、选择抽样仪,并改进现有的KSD检验质量。