Recent studies have shown that the choice of activation function can significantly affect the performance of deep learning networks. However, the benefits of novel activation functions have been inconsistent and task dependent, and therefore the rectified linear unit (ReLU) is still the most commonly used. This paper proposes a technique for customizing activation functions automatically, resulting in reliable improvements in performance. Evolutionary search is used to discover the general form of the function, and gradient descent to optimize its parameters for different parts of the network and over the learning process. Experiments with four different neural network architectures on the CIFAR-10 and CIFAR-100 image classification datasets show that this approach is effective. It discovers both general activation functions and specialized functions for different architectures, consistently improving accuracy over ReLU and other activation functions by significant margins. The approach can therefore be used as an automated optimization step in applying deep learning to new tasks.
翻译:最近的研究显示,激活功能的选择会大大影响深层学习网络的运行,然而,新激活功能的效益是不一致的,取决于任务,因此,纠正的线性单元(RELU)仍然是最常用的。本文件建议采用自动定制激活功能的技术,从而可靠地改进性能。利用进化搜索来发现该功能的一般形式,并优化其网络不同部分和学习过程的参数。在CIFAR-10和CIFAR-100图像分类数据集上进行四个不同的神经网络结构的实验表明,这一方法是有效的。它发现不同结构的一般激活功能和专业功能,不断提高RELU和其他激活功能的精度,因此,该方法可以作为一种自动优化的步骤,对新任务进行深入的学习。