We propose a new architecture for artificial neural networks called Householder-absolute neural layers, or Han-layers for short, that use Householder reflectors as weight matrices and the absolute-value function for activation. Han-layers, functioning as fully connected layers, are motivated by recent results on neural-network variability and are designed to increase activation ratio and reduce the chance of Collapse to Constants. Neural networks constructed chiefly from Han-layers are called HanNets. By construction, HanNets enjoy a theoretical guarantee that vanishing or exploding gradient never occurs. We conduct several proof-of-concept experiments. Some surprising results obtained on styled test problems suggest that, under certain conditions, HanNets exhibit an unusual ability to produce nearly perfect solutions unattainable by fully connected networks. Experiments on regression datasets show that HanNets can significantly reduce the number of model parameters while maintaining or improving the level of generalization accuracy. In addition, by adding a few Han-layers into the pre-classification FC-layer of a convolutional neural network, we are able to quickly improve a state-of-the-art result on CIFAR10 dataset. These proof-of-concept results are sufficient to necessitate further studies on HanNets to understand their capacities and limits, and to exploit their potentials in real-world applications.
翻译:我们提出一个新的人工神经网络结构,称为“家庭-固态神经层”,或“汉-层”短称“汉-层”,使用家庭式反射器作为重量矩阵和启动的绝对价值功能。汉-层作为完全连接的层的功能,受到神经-网络变化的最新结果的驱动,目的是增加激活率,减少向常数倒塌的机会。主要由汉-层建造的神经网络称为“汉-层”。通过建设,汉-网享有一种理论保证,这种理论保证永远不会消失或爆炸梯度。我们进行了几次概念校准实验。在典型测试问题上获得的一些令人惊讶的结果表明,在某些条件下,汉-层作为完全连接的层发挥完全连通性作用的功能,在产生近乎完美解决方案的能力。关于回归式数据集的实验表明,汉-网络可以大大减少模型参数的数量,同时保持或提高普遍精确度。此外,汉-网-网-梯级的预级FC-级的理论保证不会发生消失或爆炸。我们能够迅速改进系统模式应用的状态,从而在互联网上获得足够的数据极限。