Spiking neural network (SNN) operating with asynchronous discrete events shows higher energy efficiency. A popular approach to implement deep SNNs is ANN-SNN conversion combining both efficient training in ANNs and efficient inference in SNNs. However, the previous works mostly required thousands of time steps to achieve lossless conversion. In this paper, we first identify the underlying cause, i.e., misrepresentation of the negative or overflow residual membrane potential in SNNs. Furthermore, we systematically analyze the conversion error between SNNs and ANNs, and then decompose it into three folds: quantization error, clipping error, and residual membrane potential representation error. With such insights, we propose a dual-phase conversion algorithm to minimize those errors. As a result, our model achieves SOTA in both accuracy and accuracy-delay tradeoff with deep architectures (ResNet and VGG net). Specifically, we report SOTA accuracy within 16$\times$ speedup compared with the latest results. Meanwhile, lossless conversion is performed with at least 2$\times$ faster reasoning performance.
翻译:以零星离散事件运行的Spik神经网络(SNN)显示,能源效率较高。实施深层SNN的流行方法是将ANN-SNN转换成ANN-SNN,同时结合对ANNS的有效培训和对SNS的有效推断。然而,以往的工程大多需要数千个时间步骤才能实现无损转换。在本文中,我们首先找出根本原因,即对SNNS的负或溢出剩余膜潜力的误差进行误差。此外,我们系统地分析SNNS和ANNS之间的误差,然后将其分解成三个折叠:定量错误、剪切错误和残余膜潜在代表错误。我们提出一个双阶段转换算法,以最大限度地减少这些错误。结果是,我们的模型在精深结构(ResNet和VGGNet)的精度和准确度折换交易中都实现了SOTA。具体地说,我们报告SOTA准确度在16美元之内与最新结果相比的速度。同时,以至少2美元的速度进行无损的推算。