One of the reasons that many neural networks are capable of replicating complicated tasks or functions is their universality property. The past few decades have seen many attempts in providing constructive proofs for single or class of neural networks. This paper is an effort to provide a unified and constructive framework for the universality of a large class of activations including most of existing activations and beyond. At the heart of the framework is the concept of neural network approximate identity. It turns out that most of existing activations are neural network approximate identity, and thus universal in the space of continuous of functions on compacta. The framework induces several advantages. First, it is constructive with elementary means from functional analysis, probability theory, and numerical analysis. Second, it is the first unified attempt that is valid for most of existing activations. Third, as a by product, the framework provides the first university proof for some of the existing activation functions including Mish, SiLU, ELU, GELU, and etc. Fourth, it discovers new activations with guaranteed universality property. Indeed, any activation\textemdash whose $\k$th derivative, with $\k$ being an integer, is integrable and essentially bounded\textemdash is universal. Fifth, for a given activation and error tolerance, the framework provides precisely the architecture of the corresponding one-hidden neural network with predetermined number of neuron, and the values of weights/biases.
翻译:许多神经网络能够复制复杂任务或功能的原因之一是其普遍性属性。在过去几十年中,人们曾多次尝试为单一或一类神经网络提供建设性证据。本文件旨在为包括大部分现有启动活动在内的大规模类型的激活活动的普遍性提供一个统一和建设性框架。框架的核心是神经网络概念近似特性的概念。事实证明,大多数现有的激活活动是神经网络近似特性,因此在不断运行的隐约功能空间中是普遍性的。框架带来若干优势。首先,它以功能分析、概率理论和数字分析等基本手段为单一或类神经网络提供建设性证据是建设性的。第二,这是对大多数现有启动活动都有效的第一次统一尝试。第三,作为一个产品,框架为包括Mish、SilU、ELU、ELU、GELU等在内的一些现有激活功能提供了大学的第一份证据。第四,它发现新的激活活动是保证普遍性属性。事实上,任何激活活动,其第值为美元和内值的衍生值,以美元/内值为基本容忍度的架构和相应的结构框架都具有约束性。