We propose Radial Bayesian Neural Networks (BNNs): a variational approximate posterior for BNNs which scales well to large models while maintaining a distribution over weight-space with full support. Other scalable Bayesian deep learning methods, like MC dropout or deep ensembles, have discrete support-they assign zero probability to almost all of the weight-space. Unlike these discrete support methods, Radial BNNs' full support makes them suitable for use as a prior for sequential inference. In addition, they solve the conceptual challenges with the a priori implausibility of weight distributions with discrete support. The Radial BNN is motivated by avoiding a sampling problem in 'mean-field' variational inference (MFVI) caused by the so-called 'soap-bubble' pathology of multivariate Gaussians. We show that, unlike MFVI, Radial BNNs are robust to hyperparameters and can be efficiently applied to a challenging real-world medical application without needing ad-hoc tweaks and intensive tuning. In fact, in this setting Radial BNNs out-perform discrete-support methods like MC dropout. Lastly, by using Radial BNNs as a theoretically principled, robust alternative to MFVI we make significant strides in a Bayesian continual learning evaluation.
翻译:我们建议Radial Bayesian神经网络(Radial Bayesian Neal Networks ) : 一种对BNNS的变相近似近似近似近似于大模型, 在全力支持下, 将重量分布比大型模型大, 并保持在重量空间上的分布。 其他可扩缩的Bayesian深层学习方法, 如 MC 辍学或深共聚体, 具有离散支持性, 将零概率分配给几乎所有的重量空间。 与这些离散支持方法不同, Radial BNNNS 的全力支持使其适合作为先行顺序推论。 此外, 它们解决了概念上的挑战, 而在独立支持下, 重量分布比重分配的先验性不易变异性。 Radial BNNNNN是避免一个抽样问题, 也就是在不断的学习中, 以不断的双向的双向,, 将这种方法作为不断的双向,, 以不断的双向的 IM 。