We show that the output of a (residual) convolutional neural network (CNN) with an appropriate prior over the weights and biases is a Gaussian process (GP) in the limit of infinitely many convolutional filters, extending similar results for dense networks. For a CNN, the equivalent kernel can be computed exactly and, unlike "deep kernels", has very few parameters: only the hyperparameters of the original CNN. Further, we show that this kernel has two properties that allow it to be computed efficiently; the cost of evaluating the kernel for a pair of images is similar to a single forward pass through the original CNN with only one filter per layer. The kernel equivalent to a 32-layer ResNet obtains 0.84% classification error on MNIST, a new record for GPs with a comparable number of parameters.
翻译:我们显示,一个(残余的)进化神经网络(CNN)的输出在重量和偏差上具有适当的先期性,其输出是一个高斯过程(GP),在无限多的进化过滤器的限度内,为密集网络提供类似的结果。对于有线电视新闻网来说,相等的内核可以精确计算,并且与“深内核”不同,其参数非常少:只有原CNN的超参数。此外,我们显示,这个内核有两个特性,可以有效计算;对一副图像的内核进行评估的费用类似于通过原始CNN的单次前端,每层只有一个过滤器。相当于32层ResNet的内核在MNIST上获得了0.84%的分类错误,这是具有类似参数的GP的新记录。