We study norm-based uniform convergence bounds for neural networks, aiming at a tight understanding of how these are affected by the architecture and type of norm constraint, for the simple class of scalar-valued one-hidden-layer networks, and inputs bounded in Euclidean norm. We begin by proving that in general, controlling the spectral norm of the hidden layer weight matrix is insufficient to get uniform convergence guarantees (independent of the network width), while a stronger Frobenius norm control is sufficient, extending and improving on previous work. Motivated by the proof constructions, we identify and analyze two important settings where a mere spectral norm control turns out to be sufficient: First, when the network's activation functions are sufficiently smooth (with the result extending to deeper networks); and second, for certain types of convolutional networks. In the latter setting, we study how the sample complexity is additionally affected by parameters such as the amount of overlap between patches and the overall number of patches.
翻译:我们研究神经网络的基于规范的统一趋同界限,目的是深入了解这些网络如何受到结构及规范约束类型的影响,简单一类的标定值单层隐藏的网络,以及受Euclidean规范约束的投入。我们首先证明,一般而言,控制隐藏层重力矩阵的光谱规范不足以获得统一的趋同保证(独立于网络宽度),而更强大的Frobenius规范控制就足够了,比以前的工作更加广泛并有所改进。受证据构造的驱动,我们发现并分析了两个重要环境,在这两个环境中,仅仅光谱规范控制就足够了:第一,网络的激活功能足够顺利(结果扩大到更深的网络);第二,某些类型的革命性网络。在后一种环境下,我们研究抽样复杂性如何受到诸如补丁和补丁总数重叠程度等参数的额外影响。