Inspired by the diversity of biological neurons, quadratic artificial neurons can play an important role in deep learning models. The type of quadratic neurons of our interest replaces the inner-product operation in the conventional neuron with a quadratic function. Despite promising results so far achieved by networks of quadratic neurons, there are important issues not well addressed. Theoretically, the superior expressivity of a quadratic network over either a conventional network or a conventional network via quadratic activation is not fully elucidated, which makes the use of quadratic networks not well grounded. Practically, although a quadratic network can be trained via generic backpropagation, it can be subject to a higher risk of collapse than the conventional counterpart. To address these issues, we first apply the spline theory and a measure from algebraic geometry to give two theorems that demonstrate better model expressivity of a quadratic network than the conventional counterpart with or without quadratic activation. Then, we propose an effective and efficient training strategy referred to as ReLinear to stabilize the training process of a quadratic network, thereby unleashing the full potential in its associated machine learning tasks. Comprehensive experiments on popular datasets are performed to support our findings and evaluate the performance of quadratic deep learning.
翻译:在生物神经多样性的启发下,四人造神经元在深层学习模型中可以发挥重要作用。我们感兴趣的四人神经元类型以四方函数取代常规神经神经元的内产品操作。尽管四人神经元网络迄今取得了可喜的成果,但也有一些重要问题没有很好解决。理论上,四人网络在常规网络或通过四方激活的常规网络或常规网络上的超强表达性没有得到充分阐明,这使得四人网络的使用没有很好的基础。实际上,尽管我们感兴趣的四人网络可以通过一般反向分析来培训,但它可能比常规的对口网络面临更大的崩溃风险。为了解决这些问题,我们首先应用四人神经元网络的样条理论和从代数几何测算得出的测量尺度来给出两种理论,这些理论显示四人网络比传统的对口网络或无四方激活的对口网络更具有更好的示范性表达性。然后,我们提议一项高效有效的培训战略,称为ReLin, 以稳定二次网络的培训过程,但实际上,它可能会比常规的对立面网络的反向,因此,因此,我们应用了更大的崩溃风险。为了解决这些问题,我们全面学习结果,因此,因此,我们开始全面学习机器进行全面研究。