Deep learning with deep neural networks (DNNs) has attracted tremendous attention from various fields of science and technology recently. Activation functions for a DNN define the output of a neuron given an input or set of inputs. They are essential and inevitable in learning non-linear transformations and performing diverse computations among successive neuron layers. Thus, the design of activation functions is still an important topic in deep learning research. Meanwhile, theoretical studies on the approximation ability of DNNs with activation functions have been investigated within the last few years. In this paper, we propose a new activation function, named as "DLU", and investigate its approximation ability for functions with various smoothness and structures. Our theoretical results show that DLU networks can process competitive approximation performance with rational and ReLU networks, and have some advantages. Numerical experiments are conducted comparing DLU with the existing activations-ReLU, Leaky ReLU, and ELU, which illustrate the good practical performance of DLU.
翻译:与深层神经网络(DNNs)的深层学习最近吸引了科技各领域的极大关注。 DNN的激活功能定义了神经元的输出,给出了投入或一组投入。这些功能在学习非线性转变和连续神经层进行不同的计算中必不可少且不可避免。因此,激活功能的设计仍然是深层学习研究的一个重要课题。与此同时,在过去几年中,对带有激活功能的DNNs近似能力的理论研究进行了调查。在本文中,我们提议了一个新的激活功能,名为“DLU ”, 并调查其与各种平稳和结构的功能的近似能力。我们的理论结果表明, DLU 网络可以与理性和RELU 网络处理竞争性近似性表现,并具有一些优势。正在进行数值实验,将DLU 与现有的激活-RELU、Laky ReLU 和ELU ELU 进行对比,以显示DLU 的良好实际表现。