The main challenge in Bayesian models is to determine the posterior for the model parameters. Already, in models with only one or few parameters, the analytical posterior can only be determined in special settings. In Bayesian neural networks, variational inference is widely used to approximate difficult-to-compute posteriors by variational distributions. Usually, Gaussians are used as variational distributions (Gaussian-VI) which limits the quality of the approximation due to their limited flexibility. Transformation models on the other hand are flexible enough to fit any distribution. Here we present transformation model-based variational inference (TM-VI) and demonstrate that it allows to accurately approximate complex posteriors in models with one parameter and also works in a mean-field fashion for multi-parameter models like neural networks.
翻译:Bayesian 模型的主要挑战是如何确定模型参数的后方。 在仅有一个或几个参数的模型中,分析后方因素只能在特殊环境下确定。在Bayesian 神经网络中,通过变式分布,差异推导被广泛用于接近难以计算的后方因素。通常,Gaussian 被用作变式分布(Gausian-VI),由于灵活性有限,从而限制了近似质量。另一方面,变式模型具有足够的灵活性,足以适应任何分布。在这里,我们提出了基于变式模型的变异推导(TM-VI),并表明它能够精确地用一个参数模型来估计复杂的后方因素,并且也可以以中位方式为神经网络等多参数模型工作。