Spiking Neural Networks (SNNs) have attracted enormous research interest due to temporal information processing capability, low power consumption, and high biological plausibility. However, the formulation of efficient and high-performance learning algorithms for SNNs is still challenging. Most existing learning methods learn weights only, and require manual tuning of the membrane-related parameters that determine the dynamics of a single spiking neuron. These parameters are typically chosen to be the same for all neurons, which limits the diversity of neurons and thus the expressiveness of the resulting SNNs. In this paper, we take inspiration from the observation that membrane-related parameters are different across brain regions, and propose a training algorithm that is capable of learning not only the synaptic weights but also the membrane time constants of SNNs. We show that incorporating learnable membrane time constants can make the network less sensitive to initial values and can speed up learning. In addition, we reevaluate the pooling methods in SNNs and find that max-pooling will not lead to significant information loss and have the advantage of low computation cost and binary compatibility. We evaluate the proposed method for image classification tasks on both traditional static MNIST, Fashion-MNIST, CIFAR-10 datasets, and neuromorphic N-MNIST, CIFAR10-DVS, DVS128 Gesture datasets. The experiment results show that the proposed method outperforms the state-of-the-art accuracy on nearly all datasets, using fewer time-steps. Our codes are available at https://github.com/fangwei123456/Parametric-Leaky-Integrate-and-Fire-Spiking-Neuron.
翻译:由于时间信息处理能力、低电耗和高生物光度,Spik6神经网络(SNN)吸引了巨大的研究兴趣。然而,为SNN制定高效和高性能的学习算法仍然具有挑战性。大多数现有学习方法只学习重量,并且需要手工调整确定单脉冲神经神经动态的膜相关参数。这些参数通常对所有神经系统来说都是相同的,这限制了神经系统的多样性,因而也限制了由此产生的神经系统内核系统的清晰度。在本文中,我们从以下观察中得到灵感:脑区域间膜相关参数不同,并提议一种培训算法,不仅能够学习Synaptic重量,而且能够学习SNMNbrane相关参数。我们显示,可以学习的膜时间常数常数可以降低网络对初始值的敏感度,从而加速学习。此外,我们重新评价SNNNNNNNNPN的集合方法,发现与MN相联不会导致重大信息损失,而且能够了解S-RIS的拟议传统数据方法的兼容性。我们提出的数字的计算方法的优势。