Computational units in artificial neural networks follow a simplified model of biological neurons. In the biological model, the output signal of a neuron runs down the axon, splits following the many branches at its end, and passes identically to all the downward neurons of the network. Each of the downward neurons will use their copy of this signal as one of many inputs dendrites, integrate them all and fire an output, if above some threshold. In the artificial neural network, this translates to the fact that the nonlinear filtering of the signal is performed in the upward neuron, meaning that in practice the same activation is shared between all the downward neurons that use that signal as their input. Dendrites thus play a passive role. We propose a slightly more complex model for the biological neuron, where dendrites play an active role: the activation in the output of the upward neuron becomes optional, and instead the signals going through each dendrite undergo independent nonlinear filterings, before the linear combination. We implement this new model into a ReLU computational unit and discuss its biological plausibility. We compare this new computational unit with the standard one and describe it from a geometrical point of view. We provide a Keras implementation of this unit into fully connected and convolutional layers and estimate their FLOPs and weights change. We then use these layers in ResNet architectures on CIFAR-10, CIFAR-100, Imagenette, and Imagewoof, obtaining performance improvements over standard ResNets up to 1.73%. Finally, we prove a universal representation theorem for continuous functions on compact sets and show that this new unit has more representational power than its standard counterpart.
翻译:人造神经网络中的计算单位将采用简化的生物神经元模型。 在生物模型中,神经元的输出信号将沿着轴向下, 沿着许多分支的尾端分裂, 并同样传递到网络中的所有向下神经元。 每个向下神经元将使用该信号的复制件作为许多输入符之一, 将其全部整合并发布输出, 如果高于某些阈值。 在人工神经网络中, 将信号的非线性过滤器在上神经元中进行。 这意味着在实际中, 使用该信号作为输入的所有向下神经元之间共享相同的激活信号。 Dendrites因此扮演一个被动的角色。 我们为生物神经元提出一个略为复杂一点的模式, 向上神经元的输出的激活是可选的, 如果超过某些临界神经元, 则在线性组合之前, 将这个新模型应用到一个 ReLU 计算单位的计算单位, 并讨论其作为输入输入输入输入输入该信号的直径直径的直径的直径结构。 我们用这个新的计算模型进行新的计算, 直径直径直径直到直到直径的直径的直径的直径结构结构。