We study how permutation symmetries in overparameterized multi-layer neural networks generate `symmetry-induced' critical points. Assuming a network with $ L $ layers of minimal widths $ r_1^*, \ldots, r_{L-1}^* $ reaches a zero-loss minimum at $ r_1^*! \cdots r_{L-1}^*! $ isolated points that are permutations of one another, we show that adding one extra neuron to each layer is sufficient to connect all these previously discrete minima into a single manifold. For a two-layer overparameterized network of width $ r^*+ h =: m $ we explicitly describe the manifold of global minima: it consists of $ T(r^*, m) $ affine subspaces of dimension at least $ h $ that are connected to one another. For a network of width $m$, we identify the number $G(r,m)$ of affine subspaces containing only symmetry-induced critical points that are related to the critical points of a smaller network of width $r<r^*$. Via a combinatorial analysis, we derive closed-form formulas for $ T $ and $ G $ and show that the number of symmetry-induced critical subspaces dominates the number of affine subspaces forming the global minima manifold in the mildly overparameterized regime (small $ h $) and vice versa in the vastly overparameterized regime ($h \gg r^*$). Our results provide new insights into the minimization of the non-convex loss function of overparameterized neural networks.
翻译:我们研究在多层神经网络中,超参数化多层神经网络的变差对称性如何产生“对称诱导”的关键点。假设一个拥有L$层最小宽度的网络,假设一个L$层最小宽度为 r_1 ⁇,\ldots,r ⁇ L-1 ⁇ $达到零损最低值的R_1 ⁇ 美元最低值!\cdocks r ⁇ L-1 ⁇!美元偏僻点相互对调,我们发现每层多增加一个神经元足以将所有这些先前离散的微米连接成一个单一的柱形。对于一个双层超离散的网络,宽度直径直径为 r ⁇ h h=:我们明确描述全球迷你的多层:它由$(r ⁇,m)平方平方块的子空间空间至少以$连接在一起的美元为最低值。对于一个宽度的网络,我们确定一个直径直径的直径的直方形亚空间系统数为$(r,米)仅包含直径超空域值的临界点的直方形精度缩缩缩缩缩缩缩缩缩缩缩缩缩缩缩缩的系统。