Despite the rapid progress of neuromorphic computing, inadequate capacity and insufficient representation power of spiking neural networks (SNNs) severely restrict their application scope in practice. Residual learning and shortcuts have been evidenced as an important approach for training deep neural networks, but rarely did previous work assess their applicability to the characteristics of spike-based communication and spatiotemporal dynamics. In this paper, we first identify that this negligence leads to impeded information flow and accompanying degradation problem in previous residual SNNs. Then we propose a novel SNN-oriented residual block, MS-ResNet, which is able to significantly extend the depth of directly trained SNNs, e.g. up to 482 layers on CIFAR-10 and 104 layers on ImageNet, without observing any slight degradation problem. We validate the effectiveness of MS-ResNet on both frame-based and neuromorphic datasets, and MS-ResNet104 achieves a superior result of 76.02% accuracy on ImageNet, the first time in the domain of directly trained SNNs. Great energy efficiency is also observed that on average only one spike per neuron is needed to classify an input sample. We believe our powerful and scalable models will provide a strong support for further exploration of SNNs.
翻译:尽管神经畸形计算取得了迅速的进展,但神经神经网络(SNNs)能力不足和代表性力量不足,严重限制了其实际应用范围。残留学习和捷径已被证明是培训深神经网络的一个重要方法,但以往的工作很少评估其是否适用于基于悬浮的通信和神经时空动态的特点。在本文件中,我们首先指出,这一疏忽导致前残留的SNNS网络的信息流动和伴随的退化问题受到阻碍。然后,我们提出一个新的面向SNNNS的剩余区块,MS-ResNet,它能够大大扩展直接培训的SNNS的深度,例如CIFAR-10和图像网络104层的深度达482层,而没有观察任何轻微的退化问题。我们验证MS-ResNet在基于框架的和神经形态的数据集上的有效性,而MS-ResNet104在图像网络上取得了76.02 %的优异性结果,这是直接培训的SNNWs的首度领域。高能效还观察到,平均只需要每神经型的1个坚挺的峰,才能对输入样本进行进一步分类。我们相信SNNNS的强大和卡将进一步提供支持。