Sampling from an unnormalized probability distribution is a fundamental problem in machine learning with applications including Bayesian modeling, latent factor inference, and energy-based model training. After decades of research, variations of MCMC remain the default approach to sampling despite slow convergence. Auxiliary neural models can learn to speed up MCMC, but the overhead for training the extra model can be prohibitive. We propose a fundamentally different approach to this problem via a new Hamiltonian dynamics with a non-Newtonian momentum. In contrast to MCMC approaches like Hamiltonian Monte Carlo, no stochastic step is required. Instead, the proposed deterministic dynamics in an extended state space exactly sample the target distribution, specified by an energy function, under an assumption of ergodicity. Alternatively, the dynamics can be interpreted as a normalizing flow that samples a specified energy model without training. The proposed Energy Sampling Hamiltonian (ESH) dynamics have a simple form that can be solved with existing ODE solvers, but we derive a specialized solver that exhibits much better performance. ESH dynamics converge faster than their MCMC competitors enabling faster, more stable training of neural network energy models.
翻译:从未实现的概率分布中取样,是机器学习包括贝耶斯模型模型、潜在因素推断和能源模型培训在内的应用的根本性问题。经过数十年的研究后,MCMC的变异仍然是抽样的默认方法,尽管速度趋同缓慢。辅助神经模型可以学会加速MCMC,但额外模型的培训管理费用可能令人望而却步。我们提议通过一个新的汉密尔顿动力和非纽顿动力的新的汉密尔顿动力来从根本上改变这一问题。与汉密尔顿蒙特卡洛这样的MCMC方法相比,不需要采取软体步骤。相反,在扩展的状态空间中,拟议的确定性动态确切地对由能源函数指定的目标分布进行抽样。在一种假设中,一种能源函数的假设下,这种变异性可以被解释为一种正常的流,在没有培训的情况下对特定能源模型进行抽样。拟议的能源采集汉密尔密尔顿(ESH)动态有一个简单的形式,可以与现有的内存解解剂一起解决,但我们可以找到一个表现更好的专门解答器。ESH的动态比它们的MCMC竞争者能够更快、更稳定地进行神经网络模型的培训。