We study the optimization landscape and the stability properties of training problems with squared loss for neural networks and general nonlinear conic approximation schemes. It is demonstrated that, if a nonlinear conic approximation scheme is considered that is (in an appropriately defined sense) more expressive than a classical linear approximation approach and if there exist unrealizable label vectors, then a training problem with squared loss is necessarily unstable in the sense that its solution set depends discontinuously on the label vector in the training data. We further prove that the same effects that are responsible for these instability properties are also the reason for the emergence of saddle points and spurious local minima, which may be arbitrarily far away from global solutions, and that neither the instability of the training problem nor the existence of spurious local minima can, in general, be overcome by adding a regularization term to the objective function that penalizes the size of the parameters in the approximation scheme. The latter results are shown to be true regardless of whether the assumption of realizability is satisfied or not. We demonstrate that our analysis in particular applies to training problems for free-knot interpolation schemes and deep and shallow neural networks with variable widths that involve an arbitrary mixture of various activation functions (e.g., binary, sigmoid, tanh, arctan, soft-sign, ISRU, soft-clip, SQNL, ReLU, leaky ReLU, soft-plus, bent identity, SILU, ISRLU, and ELU). In summary, the findings of this paper illustrate that the improved approximation properties of neural networks and general nonlinear conic approximation instruments are linked in a direct and quantifiable way to undesirable properties of the optimization problems that have to be solved in order to train them.
翻译:我们研究的是神经网络和一般非线性近似计划的平方损失的培训问题的最佳景观和稳定性特征。我们进一步证明,造成这些不稳定性效应的同样影响也是出现马鞍点和虚假的本地迷你现象的原因,而这些地方迷你现象可能与全球解决方案相去甚远,如果认为非线性近似方案比古典线性近似方法(在适当定义的意义上)更具有显眼性,而且如果存在无法实现的标签矢量,那么,如果存在平方损失的培训问题必然是不稳定的,因为其解决办法取决于培训数据中的标签矢量。我们进一步证明,造成这些不稳定性能的同样影响也是出现马鞍点和虚假的当地迷你现象的原因。如果认为非线性近似方案(在定义明确的意义上)比古典性近似近似近似近似方法更为明显,培训问题的不稳定性或虚伪的本地迷你近似近似方法,一般纸质变软性纸质的内质、软性纸质的内质的内质、内质的内质、内质的内质、内质、内质的内质的内质性、内质、内质、内质的内质、内质的内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质性、内质、内质、内质性、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质的、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、内质、