Spiking neural networks (SNN) are usually more energy-efficient as compared to Artificial neural networks (ANN), and the way they work has a great similarity with our brain. Back-propagation (BP) has shown its strong power in training ANN in recent years. However, since spike behavior is non-differentiable, BP cannot be applied to SNN directly. Although prior works demonstrated several ways to approximate the BP-gradient in both spatial and temporal directions either through surrogate gradient or randomness, they omitted the temporal dependency introduced by the reset mechanism between each step. In this article, we target on theoretical completion and investigate the effect of the missing term thoroughly. By adding the temporal dependency of the reset mechanism, the new algorithm is more robust to learning-rate adjustments on a toy dataset but does not show much improvement on larger learning tasks like CIFAR-10. Empirically speaking, the benefits of the missing term are not worth the additional computational overhead. In many cases, the missing term can be ignored.
翻译:与人工神经网络(ANN)相比,螺旋神经网络(SNN)通常更具有能源效率,而且它们的工作方式与我们的大脑非常相似。后推进(BP)近年来在培训ANN方面表现出强大的力量。然而,由于峰值行为是不可区分的,因此不能直接对SNN适用BP。虽然先前的工作表明通过代位梯度或随机性在空间和时间方向上接近BP,但是它们忽略了每个步骤之间重新设置机制引入的时间依赖性。在本篇文章中,我们的目标是理论完成并彻底调查缺失术语的效果。由于增加了重设机制的时间依赖性,新的算法对于学习对微量数据集的调整更为有力,但对于像CIFAR-10这样的较大学习任务没有多大改进。 具有讽刺意味的是,缺失术语的好处不值得额外的计算性间接损失。在许多情况下,缺失的术语可以被忽略。