We study the extent to which wide neural networks may be approximated by Gaussian processes when initialized with random weights. It is a well-established fact that as the width of a network goes to infinity, its law converges to that of a Gaussian process. We make this quantitative by establishing explicit convergence rates for the central limit theorem in an infinite-dimensional functional space, metrized with a natural transportation distance. We identify two regimes of interest; when the activation function is polynomial, its degree determines the rate of convergence, while for non-polynomial activations, the rate is governed by the smoothness of the function.
翻译:我们研究高山进程在以随机重量初始化时,大神经网络可能近似于高山进程的程度。一个公认的事实是,当网络的宽度达到无限时,其法律与高山进程的法律趋于一致。我们通过在无限的功能空间为中心极限定出明确的趋同率,与自然迁移距离相匹配,来进行这种定量。我们确定了两种感兴趣的制度;当激活功能是多元的时,其程度决定了趋同率,而对于非球体激活而言,其速度则取决于功能的平稳性。