Many feedforward neural networks generate continuous and piecewise-linear (CPWL) mappings. Specifically, they partition the input domain into regions on which the mapping is an affine function. The number of these so-called linear regions offers a natural metric to characterize the expressiveness of CPWL mappings. Although the precise determination of this quantity is often out of reach, bounds have been proposed for specific architectures, including the well-known ReLU and Maxout networks. In this work, we propose a more general perspective and provide precise bounds on the maximal number of linear regions of CPWL networks based on three sources of expressiveness: depth, width, and activation complexity. Our estimates rely on the combinatorial structure of convex partitions and highlight the distinctive role of depth which, on its own, is able to exponentially increase the number of regions. We then introduce a complementary stochastic framework to estimate the average number of linear regions produced by a CPWL network architecture. Under reasonable assumptions, the expected density of linear regions along any 1D path is bounded by the product of depth, width, and a measure of activation complexity (up to a scaling factor). This yields an identical role to the three sources of expressiveness: no exponential growth with depth is observed anymore.
翻译:许多进料神经网络生成连续和小线线(CPWL)映射。 具体地说, 这些所谓的线性区域的数量为CPWL映射的清晰度提供了自然的度量。 虽然这一数量的精确度往往无法达到,但已经为特定结构提出了界限, 包括众所周知的ReLU和Maxout网络。 在这项工作中, 我们提出了一个更宽泛的视角, 并提供了CPWL网络最大线性区域的最大数量线性界限, 其依据是三种表达性来源: 深度、 宽度和激活复杂性。 这些所谓的线性区域的数量提供了一种自然度指标, 以描述CPWL网络的清晰度特征。 我们的估计数取决于convex分区的组合结构, 并突出其独特的深度作用, 而这种作用本身能够指数性地增加区域的数量。 然后我们引入一个互补的随机框架, 以估计CPLL网络结构产生的线性区域的平均数量。 在合理的假设下, 沿任何一条方向线性区域的预期密度将受到深度、 深度、 宽度和激活复杂性的三种因素的约束。 我们所观察到的精确度作用是深度、 的指数性系数的产值。