Spiking and Quantized Neural Networks (NNs) are becoming exceedingly important for hyper-efficient implementations of Deep Learning (DL) algorithms. However, these networks face challenges when trained using error backpropagation, due to the absence of gradient signals when applying hard thresholds. The broadly accepted trick to overcoming this is through the use of biased gradient estimators: surrogate gradients which approximate thresholding in Spiking Neural Networks (SNNs), and Straight-Through Estimators (STEs), which completely bypass thresholding in Quantized Neural Networks (QNNs). While noisy gradient feedback has enabled reasonable performance on simple supervised learning tasks, it is thought that such noise increases the difficulty of finding optima in loss landscapes, especially during the later stages of optimization. By periodically boosting the Learning Rate (LR) during training, we expect the network can navigate unexplored solution spaces that would otherwise be difficult to reach due to local minima, barriers, or flat surfaces. This paper presents a systematic evaluation of a cosine-annealed LR schedule coupled with weight-independent adaptive moment estimation as applied to Quantized SNNs (QSNNs). We provide a rigorous empirical evaluation of this technique on high precision and 4-bit quantized SNNs across three datasets, demonstrating (close to) state-of-the-art performance on the more complex datasets. Our source code is available at this link: https://github.com/jeshraghian/QSNNs.
翻译:Spiking和量化神经网络(NNS)对于高效益实施深层学习(DL)算法(QNNS)正在变得极为重要。 然而,这些网络在培训时使用错误反反演法时面临挑战,原因是在应用硬阈时没有梯度信号。 克服这一点的广泛接受的伎俩是使用偏差梯度估计器:在Spiking神经网络(SNNS)中近似临界值的代谢梯度,以及直流的复杂模拟器(STEs),它们完全绕过量化神经网络(QNNNS)的临界值。虽然由于噪音的梯度反馈使得在简单监管的学习任务中能够合理表现,但人们认为这种噪音增加了在损失地貌中找到opima的难度,特别是在后期优化阶段。通过在培训期间定期提高学习率,我们期望网络能够浏览由于本地微型、屏障或平坦表面而难以达到的未爆炸式解决方案空间。 这份论文展示了Concial-anal Q-ereal 精确度(Snal-deal cloveal dal dalalalalalalalalalal) 将S-salalalalal deviews to wetraviews 用于Syalds 4saldaldaldaldaldaldaldaldaldaldaldaldaldaldaldaldaldaldals) 在S-salsaldaldaldal