Recently, particle-based variational inference (ParVI) methods have gained interest because they directly minimize the Kullback-Leibler divergence and do not suffer from approximation errors from the evidence-based lower bound. However, many ParVI approaches do not allow arbitrary sampling from the posterior, and the few that do allow such sampling suffer from suboptimality. This work proposes a new method for learning to approximately sample from the posterior distribution. We construct a neural sampler that is trained with the functional gradient of the KL-divergence between the empirical sampling distribution and the target distribution, assuming the gradient resides within a reproducing kernel Hilbert space. Our generative ParVI (GPVI) approach maintains the asymptotic performance of ParVI methods while offering the flexibility of a generative sampler. Through carefully constructed experiments, we show that GPVI outperforms previous generative ParVI methods such as amortized SVGD, and is competitive with ParVI as well as gold-standard approaches like Hamiltonian Monte Carlo for fitting both exactly known and intractable target distributions.
翻译:最近,基于粒子的变异推断方法(ParVI)引起了人们的兴趣,因为它们直接最大限度地缩小了Kullback-Leverer的偏差,而且不会因基于证据的较低约束而出现近似差差。然而,许多ParVI方法不允许从远地点任意采样,而能够进行这种采样的少数少数方法则存在次优性。这项工作提出了一种新的方法,用于学习从远地点分布中大约采样的方法。我们建造了一个神经取样器,在实验采样分布与目标分布之间用KL-diverence的功能梯度进行训练,假设梯度位于再生产内核圈Hilbert空间之内。我们的基因化ParVI(GPVI)方法保持了ParVI方法的无症状性能,同时提供了基因化采样器的灵活性。我们通过精心构建的实验,表明GPVI优于先前的基因化PAVI方法,如摊销SVGD,并且与ParVI具有竞争力,而且与金质标准方法如汉密尔顿·蒙特卡洛(Carlo)一样,以适应精确和棘手的目标分布。