Despite the rapid progress of neuromorphic computing, inadequate capacity and insufficient representation power of spiking neural networks (SNNs) severely restrict their application scope in practice. Residual learning and shortcuts have been evidenced as an important approach for training deep neural networks, but rarely did previous work assess their applicability to the characteristics of spike-based communication and spatiotemporal dynamics. In this paper, we first identify that this negligence leads to impeded information flow and the accompanying degradation problem in previous residual SNNs. To address this issue, we propose a novel SNN-oriented residual architecture termed MS-ResNet, which establishes membrane-based shortcut pathways, and further prove that the gradient norm equality can be achieved in MS-ResNet by introducing block dynamical isometry theory, which ensures the network can be well-behaved in a depth-insensitive way. Thus we are able to significantly extend the depth of directly trained SNNs, e.g., up to 482 layers on CIFAR-10 and 104 layers on ImageNet, without observing any slight degradation problem. To validate the effectiveness of MS-ResNet, experiments on both frame-based and neuromorphic datasets are conducted. MS-ResNet104 achieves a superior result of 76.02% accuracy on ImageNet, which is the highest to our best knowledge in the domain of directly trained SNNs. Great energy efficiency is also observed, with an average of only one spike per neuron needed to classify an input sample. We believe our powerful and scalable models will provide a strong support for further exploration of SNNs.
翻译:尽管神经变形计算取得了快速进展,但神经神经网络(SNN)能力不足和代表性力量不足,严重限制了其实际应用范围。残余学习和捷径被证明是培训深神经网络的一个重要方法,但很少前期工作评估其是否适用于基于螺旋的通信和空间时空动态的特点。在本文件中,我们首先发现,这一疏忽导致阻碍信息流动,并随之导致先前遗留的SNNN的退化问题。为了解决这一问题,我们提议建立一个名为MS-ResNet的面向S-ResNet的新的SNND剩余结构,建立基于膜的捷径通道,并进一步证明在MS-ResNet中可以实现梯度规范平等,为此引入了块状动态等量理论,确保网络能够以深度不敏感的方式运行。因此,我们能够大大扩展直接培训的SNNNW的深度,例如,在CIFAR-10和104层图像网络上,在不观察到任何轻微的退化问题。为了验证MS-ResNet的有效性,在S-ResNet中,一个基于框架和S-rgrodeal的精度精度的S-alalalalal 10的SNet实验, 也是我们所观测到的S-ral-ralalalal-al-al-al-al-al-al-al-deal-al-al-al-al-al-al-als a a 需要一个我们一个最精度的Smaxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx的精度,我们的精度,我们测的精度,我们S-x的精度的精度</s>