Recent empirical work has shown that hierarchical convolutional kernels inspired by convolutional neural networks (CNNs) significantly improve the performance of kernel methods in image classification tasks. A widely accepted explanation for their success is that these architectures encode hypothesis classes that are suitable for natural images. However, understanding the precise interplay between approximation and generalization in convolutional architectures remains a challenge. In this paper, we consider the stylized setting of covariates (image pixels) uniformly distributed on the hypercube, and characterize exactly the RKHS of kernels composed of single layers of convolution, pooling, and downsampling operations. We use this characterization to compute sharp asymptotics of the generalization error for any given function in high-dimension. In particular, we quantify the gain in sample complexity brought by enforcing locality with the convolution operation and approximate translation invariance with average pooling. Notably, these results provide a precise description of how convolution and pooling operations trade off approximation with generalization power in one layer convolutional kernels.
翻译:最近的实证工作表明,受进化神经网络(CNNs)启发的级级共振内核显著改进了内核方法在图像分类任务中的性能。对其成功的一个普遍接受的解释是,这些结构将适合自然图像的假设类别编码成。然而,了解进化结构中近似和一般化之间的精确相互作用仍然是一个挑战。在本文中,我们认为在超立体上统一分布的共变形(象素)的元体设置,并准确地描述由单层共振、集合和降压操作构成的内核内核的RKHS。我们用这种特征来计算任何高变形功能中一般化错误的精度。特别是,我们用数量来说明通过实施进化操作和平均集化的近似变异性带来的样本复杂性的增益。值得注意的是,这些结果准确地描述了在一层共振内核内核中与一般化力量的近似交易是如何进行的。