This work studies approximation based on single-hidden-layer feedforward and recurrent neural networks with randomly generated internal weights. These methods, in which only the last layer of weights and a few hyperparameters are optimized, have been successfully applied in a wide range of static and dynamic learning problems. Despite the popularity of this approach in empirical tasks, important theoretical questions regarding the relation between the unknown function, the weight distribution, and the approximation rate have remained open. In this work it is proved that, as long as the unknown function, functional, or dynamical system is sufficiently regular, it is possible to draw the internal weights of the random (recurrent) neural network from a generic distribution (not depending on the unknown object) and quantify the error in terms of the number of neurons and the hyperparameters. In particular, this proves that echo state networks with randomly generated weights are capable of approximating a wide class of dynamical systems arbitrarily well and thus provides the first mathematical explanation for their empirically observed success at learning dynamical systems.
翻译:这份基于单隐藏层进料和经常神经网络的近似工作研究基于单隐藏层进料和随机生成内部重量的神经网络,这些方法仅优化了最后一层重量和几个超参数,成功地应用于一系列静态和动态学习问题。尽管这种方法在实证工作中很受欢迎,但关于未知函数、重量分布和近似率之间关系的重要理论问题仍然开放。在这项工作中,可以证明,只要未知函数、功能或动态系统足够正常,就有可能从一般分布(不取决于未知对象)中提取随机(经常)神经网络的内部重量,并用神经数量和超光量参数来量化错误。特别是,这证明随机生成重量的回声状态网络能够任意地适应广泛的动态系统类别,从而为其在学习动态系统方面观察到的成功经验提供第一个数学解释。