In this article we consider Bayesian inference associated to deep neural networks (DNNs) and in particular, trace-class neural network (TNN) priors which were proposed by Sell et al. [39]. Such priors were developed as more robust alternatives to classical architectures in the context of inference problems. For this work we develop multilevel Monte Carlo (MLMC) methods for such models. MLMC is a popular variance reduction technique, with particular applications in Bayesian statistics and uncertainty quantification. We show how a particular advanced MLMC method that was introduced in [4] can be applied to Bayesian inference from DNNs and establish mathematically, that the computational cost to achieve a particular mean square error, associated to posterior expectation computation, can be reduced by several orders, versus more conventional techniques. To verify such results we provide numerous numerical experiments on model problems arising in machine learning. These include Bayesian regression, as well as Bayesian classification and reinforcement learning.
翻译:在本文中,我们考虑了与深神经网络(DNNs)相关的贝氏推论,特别是Sell等人(39)提议的微级神经网络(TNN)前科,这些前科是作为古典建筑在推理问题背景下的更强有力的替代物而开发的。我们为这种模型开发了多层次的蒙特卡洛(MLMC)方法。MLMC是一种减少差异的流行技术,在巴伊西亚统计和不确定性量化中特别应用了这种技术。我们展示了如何将[4]中引入的超高级MLMC方法应用于从DNS(4)中推断出的巴伊西亚神经网络,并用数学方法确定,实现与后视计算相关的特定中度平方错误的计算成本可以通过几个订单,而不是更多的常规技术加以降低。为了核实这些结果,我们提供了许多关于机器学习中出现的模型问题的数值实验。其中包括贝伊斯回归,以及巴伊西亚分类和强化学习。