To better understand the theoretical behavior of large neural networks, several works have analyzed the case where a network's width tends to infinity. In this regime, the effect of random initialization and the process of training a neural network can be formally expressed with analytical tools like Gaussian processes and neural tangent kernels. In this paper, we review methods for quantifying uncertainty in such infinite-width neural networks and compare their relationship to Gaussian processes in the Bayesian inference framework. We make use of several equivalence results along the way to obtain exact closed-form solutions for predictive uncertainty.
翻译:为了更好地了解大型神经网络的理论行为,一些作品分析了网络宽度倾向于无限的情况。 在这种制度下,随机初始化的影响和神经网络培训过程可以用高山过程和神经相近内核等分析工具正式表达。 在本文中,我们审查如何量化这种无限宽神经网络的不确定性,并比较它们在巴伊西亚推理框架中与高斯进程的关系。我们沿途径利用几种等值结果获得准确的封闭式的预测不确定性解决方案。