The universal approximation property of various machine learning models is currently only understood on a case-by-case basis, limiting the rapid development of new theoretically justified neural network architectures and blurring our understanding of our current models' potential. This paper works towards overcoming these challenges by presenting a characterization, a representation, a construction method, and an existence result, each of which applies to any universal approximator on most function spaces of practical interest. Our characterization result is used to describe which activation functions allow the feed-forward architecture to maintain its universal approximation capabilities when multiple constraints are imposed on its final layers and its remaining layers are only sparsely connected. These include a rescaled and shifted Leaky ReLU activation function but not the ReLU activation function. Our construction and representation result is used to exhibit a simple modification of the feed-forward architecture, which can approximate any continuous function with non-pathological growth, uniformly on the entire Euclidean input space. This improves the known capabilities of the feed-forward architecture.
翻译:各种机器学习模型的普遍近似特性目前只能逐案理解,限制了新理论上合理的神经网络结构的快速发展,模糊了我们对当前模型潜力的理解。本文件致力于通过提供特征描述、描述、构建方法和存在结果来克服这些挑战,每个特征描述、描述、构建方法和存在结果都适用于任何具有实际利益的大多数功能空间的通用近近似特性。我们的定性结果用于描述哪些激活功能允许进料前方结构在对最终层施加多重限制时保持其普遍近似能力,而其余层则仅与稀疏连接。其中包括重定和转移的LAKY ReLU启动功能,而不是RELU启动功能。我们的构建和代表结果被用来展示饲料前方结构的简单修改,这种修改可以与非病态增长相近,与整个Euclidean输入空间一致。这提高了已知的进料前方结构的能力。