Spiking Neural Networks (SNNs) have attracted enormous research interest due to temporal information processing capability, low power consumption, and high biological plausibility. However, the formulation of efficient and high-performance learning algorithms for SNNs is still challenging. Most existing learning methods learn the synaptic-related parameters only, and require manual tuning of the membrane-related parameters that determine the dynamics of single spiking neurons. These parameters are typically chosen to be the same for all neurons, which limits the diversity of neurons and thus the expressiveness of the resulting SNNs. In this paper, we take inspiration from the observation that membrane-related parameters are different across brain regions, and propose a training algorithm that is capable to learn not only the synaptic weights but also the membrane time constants of SNN. We show that incorporating learnable membrane time constants can make the network less sensitive to initial values and can speed up learning. In addition, we reevaluate the pooling methods in SNNs and find that max-pooling is able to increase the fitting capacity of SNNs in temporal tasks, as well as reduce the computation cost. We evaluate the proposed method for image classification tasks on both traditional static MNIST, Fashion-MNIST, CIFAR-10 datasets, and neuromorphic N-MNIST, CIFAR10-DVS, DVS128 Gesture datasets. The experiment results show that the proposed method outperforms the state-of-the-art accuracy on nearly all datasets, using fewer time-steps.
翻译:由于时间信息处理能力、低电耗和高生物光度,Spik Neal网络(SNN)吸引了巨大的研究兴趣。然而,为SNN制定高效和高性能的学习算法仍然具有挑战性。大多数现有学习方法只学习与合成相关的参数,并且需要人工调整确定单弹跳神经元动态的膜相关参数。这些参数通常对所有神经系统来说都是相同的,这限制了神经系统的多样性,因而也限制了由此产生的SNNNNN的外观性。在本文中,我们从以下观察中汲取灵感:脑神经系统与膜有关的参数不同,并提议一种培训算法,不仅能够学习合成值相关参数,而且能够学习SNNNNNN的膜相关参数。我们显示,将可学习的膜时间常数常数纳入网络可以降低初始值的敏感度,从而加快学习速度。此外,我们重新评价SNNNNPN的集合方法,并发现与MER相关的参数不同,我们提议在SNNMR的S-R的S-S-S-S-S-S-S-S-S-S-S-IL-S-S-S-S-S-S-S-S-S-S-S-IAR-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-SL-S-S-S-S-SL-S-S-S-SL-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-SL-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S