Markov Chain Monte Carlo (MCMC) methods are a powerful tool for computation with complex probability distributions. However the performance of such methods is critically dependant on properly tuned parameters, most of which are difficult if not impossible to know a priori for a given target distribution. Adaptive MCMC methods aim to address this by allowing the parameters to be updated during sampling based on previous samples from the chain at the expense of requiring a new theoretical analysis to ensure convergence. In this work we extend the convergence theory of adaptive MCMC methods to a new class of methods built on a powerful class of parametric density estimators known as normalizing flows. In particular, we consider an independent Metropolis-Hastings sampler where the proposal distribution is represented by a normalizing flow whose parameters are updated using stochastic gradient descent. We explore the practical performance of this procedure on both synthetic settings and in the analysis of a physical field system and compare it against both adaptive and non-adaptive MCMC methods.
翻译:Markov链条蒙特卡洛(MCMC)方法是计算概率分布复杂情况的有力工具,但这种方法的性能严重依赖适当调整的参数,其中多数即使并非不可能,也很难知道某一目标分布的先验性。适应性MC方法的目的是解决这一问题,允许在根据先前的链条样本取样时更新参数,而牺牲了要求新的理论分析以确保趋同。在这项工作中,我们将适应性MCMC方法的趋同理论理论推广到建立在被称为正常流动的强力参数密度估计器的新型方法上。我们特别认为,在建议分布以正常流为代表的独立的Metropolis-Hastings采样器,其参数利用随机梯度梯度梯度下降加以更新。我们探讨这一程序在合成环境以及在对物理场系统的分析中的实际表现,并与适应性和非适应性MCM方法进行比较。