Deep Spiking Neural Networks (SNNs) present optimization difficulties for gradient-based approaches due to discrete binary activation and complex spatial-temporal dynamics. Considering the huge success of ResNet in deep learning, it would be natural to train deep SNNs with residual learning. Previous Spiking ResNet mimics the standard residual block in ANNs and simply replaces ReLU activation layers with spiking neurons, which suffers the degradation problem and can hardly implement residual learning. In this paper, we propose the spike-element-wise (SEW) ResNet to realize residual learning in deep SNNs. We prove that the SEW ResNet can easily implement identity mapping and overcome the vanishing/exploding gradient problems of Spiking ResNet. We evaluate our SEW ResNet on ImageNet and DVS Gesture datasets, and show that SEW ResNet outperforms the state-of-the-art directly trained SNNs in both accuracy and time-steps. Moreover, SEW ResNet can achieve higher performance by simply adding more layers, providing a simple method to train deep SNNs. To our best knowledge, this is the first time that directly training deep SNNs with more than 100 layers becomes possible.
翻译:深 Spiking神经网络(SNNS) 显示,由于离散的二进制激活和复杂的空间时空动态,基于梯度的方法在优化方面遇到困难。 考虑到ResNet在深层学习中的巨大成功,对深层SNNS进行培训是自然的。 先前的Spiking ResNet模仿了非NNS的标准残留块,而只是用弹出神经元取代了RELU的启动层,这造成了退化问题,也难以实施剩余学习。 在本文中,我们建议通过尖刺元素ResNet实现深层SNNS的剩余学习。 我们证明, SEW ResNet可以轻松地进行身份绘图,克服Spiking ResNet的消失/爆破梯度问题。 我们评估我们的图像网络和 DVS Gesture数据集上的SEW ResNet, 显示SEW ResNet在准确性和时间步骤上都超越了经过直接培训的状态。此外, SEWResNet可以实现更高的业绩,只需增加更多的层次, 提供比深层 SNN 更深层次的训练最简单的方法。