The simulation of human neurons and neurotransmission mechanisms has been realized in deep neural networks based on the theoretical implementations of activation functions. However, recent studies have reported that the threshold potential of neurons exhibits different values according to the locations and types of individual neurons, and that the activation functions have limitations in terms of representing this variability. Therefore, this study proposes a simple yet effective activation function that facilitates different thresholds and adaptive activations according to the positions of units and the contexts of inputs. Furthermore, the proposed activation function mathematically exhibits a more generalized form of Swish activation function, and thus we denoted it as Adaptive SwisH (ASH). ASH highlights informative features that exhibit large values in the top percentiles in an input, whereas it rectifies low values. Most importantly, ASH exhibits trainable, adaptive, and context-aware properties compared to other activation functions. Furthermore, ASH represents general formula of the previously studied activation function and provides a reasonable mathematical background for the superior performance. To validate the effectiveness and robustness of ASH, we implemented ASH into many deep learning models for various tasks, including classification, detection, segmentation, and image generation. Experimental analysis demonstrates that our activation function can provide the benefits of more accurate prediction and earlier convergence in many deep learning applications.
翻译:人类神经元和神经传输机制的模拟是在激活功能的理论实施基础上在深神经网络中实现的,然而,最近的研究表明,神经元的临界值潜力根据各个神经元的位置和类型呈现不同值,激活功能在代表这种变异方面有局限性。因此,本研究报告提出一个简单而有效的激活功能,根据单位的位置和投入的背景,促进不同的阈值和适应性激活。此外,拟议的激活功能数学上展示了一种更加普遍的苏氏活化功能形式,因此我们把它称为适应性SWISH(ASH)。ASH突出显示在输入的顶部百分位中显示出巨大价值的信息特征,而它又重新定位低值。最重要的是,ASH展览可培训、适应性和环境觉悟特性与其他激活功能相比是有限的。此外,ASH代表了先前研究的激活功能的一般公式,并为优异性表现提供了合理的数学背景。为了验证ASH的实效和稳健性,我们把ASHASH应用成许多深层次学习模型,包括分类、检测、分化、分化和图像生成中的更精确性化功能。实验性能分析能够展示我们早期的早期的学习功能。