Spiking neural networks (SNNs) are promising brain-inspired energy-efficient models. Recent progress in training methods has enabled successful deep SNNs on large-scale tasks with low latency. Particularly, backpropagation through time (BPTT) with surrogate gradients (SG) is popularly used to achieve high performance in a very small number of time steps. However, it is at the cost of large memory consumption for training, lack of theoretical clarity for optimization, and inconsistency with the online property of biological learning and rules on neuromorphic hardware. Other works connect spike representations of SNNs with equivalent artificial neural network formulation and train SNNs by gradients from equivalent mappings to ensure descent directions. But they fail to achieve low latency and are also not online. In this work, we propose online training through time (OTTT) for SNNs, which is derived from BPTT to enable forward-in-time learning by tracking presynaptic activities and leveraging instantaneous loss and gradients. Meanwhile, we theoretically analyze and prove that gradients of OTTT can provide a similar descent direction for optimization as gradients based on spike representations under both feedforward and recurrent conditions. OTTT only requires constant training memory costs agnostic to time steps, avoiding the significant memory costs of BPTT for GPU training. Furthermore, the update rule of OTTT is in the form of three-factor Hebbian learning, which could pave a path for online on-chip learning. With OTTT, it is the first time that two mainstream supervised SNN training methods, BPTT with SG and spike representation-based training, are connected, and meanwhile in a biologically plausible form. Experiments on CIFAR-10, CIFAR-100, ImageNet, and CIFAR10-DVS demonstrate the superior performance of our method on large-scale static and neuromorphic datasets in small time steps.
翻译:Spik 神经神经网络(SNNS)是充满希望的由大脑启发的智能智能模型。最近培训方法的进展使得在低潜潜伏的大型任务中能够成功地进行深潜 SNNS。 特别是,通过代用梯度(SG)在时间上进行反向调整(BBTTT)被广泛用来在非常小的时间步骤中达到高性能。 但是,这是大量用于培训的记忆消耗、缺乏优化的理论清晰度、与生物学习的在线属性和神经变异硬件规则不一致的代价。 其他工作将SNNNS的峰值表示与同等的人工神经网络配制和从同等的绘图中培训SNNNNTS。 但是,他们未能达到低潜伏的GTTT(G)反向反向反向反向反向反向反向反向反向的反向反向回推(OTF) 。我们提议,从BPTTT系统追踪前级活动以及利用瞬间损失和梯变梯。我们从理论上分析并证明,TTTT的梯度的梯度显示其向向向向向向向后向向后向向向最近向向方向的学习需要不断更新的学习成本成本成本。