We introduce a new notion of complexity of functions and we show that it has the following properties: (i) it governs a PAC Bayes-like generalization bound, (ii) for neural networks it relates to natural notions of complexity of functions (such as the variation), and (iii) it explains the generalization gap between neural networks and linear schemes. While there is a large set of papers which describes bounds that have each such property in isolation, and even some that have two, as far as we know, this is a first notion that satisfies all three of them. Moreover, in contrast to previous works, our notion naturally generalizes to neural networks with several layers. Even though the computation of our complexity is nontrivial in general, an upper-bound is often easy to derive, even for higher number of layers and functions with structure, such as period functions. An upper-bound we derive allows to show a separation in the number of samples needed for good generalization between 2 and 4-layer neural networks for periodic functions.
翻译:我们引入了功能复杂性的新概念,并表明它具有以下属性:(一)它管辖PAC Bayes式的通用约束,(二)它适用于神经网络,它涉及功能复杂性的自然概念(例如变异),以及(三)它解释了神经网络和线性计划之间的一般差距。虽然有一大套文件,描述了每个这些属性都与世隔绝的界限,而且就我们所知,甚至有些有两种属性,但这是第一个满足所有三个属性的概念。此外,与以往的著作不同,我们的概念自然地将神经网络分为几个层次。尽管我们的复杂性的计算一般是非边际的,但一个上限往往很容易产生,即使具有较高的层次和结构的功能,例如时段功能。我们产生的上层可以显示为定期功能在2至4级的神经网络之间进行良好综合所需的样本数量。</s>