The machine learning community has become increasingly interested in the energy efficiency of neural networks. The Spiking Neural Network (SNN) is a promising approach to energy-efficient computing, since its activation levels are quantized into temporally sparse, one-bit values (i.e., "spike" events), which additionally converts the sum over weight-activity products into a simple addition of weights (one weight for each spike). However, the goal of maintaining state-of-the-art (SotA) accuracy when converting a non-spiking network into an SNN has remained an elusive challenge, primarily due to spikes having only a single bit of precision. Adopting tools from signal processing, we cast neural activation functions as quantizers with temporally-diffused error, and then train networks while smoothly interpolating between the non-spiking and spiking regimes. We apply this technique to the Legendre Memory Unit (LMU) to obtain the first known example of a hybrid SNN outperforming SotA recurrent architectures -- including the LSTM, GRU, and NRU -- in accuracy, while reducing activities to at most 3.74 bits on average with 1.26 significant bits multiplying each weight. We discuss how these methods can significantly improve the energy efficiency of neural networks.
翻译:机器学习界对神经网络的能源效率越来越感兴趣。 Spiking神经网络(SNN)是节能计算的一个很有希望的方法,因为其激活水平被量化成暂时稀少的一比特值(即“spike”事件),从而将超重活动产品的总和转换成简单的重量加增(每个重量加增一个重量)。然而,在将非喷射网络转换成SNNN时,保持最新状态(Sota)准确性的目标仍然是一个难以捉摸的挑战,主要因为峰值只有一丁点精确度。从信号处理中采用工具,我们将神经激活功能作为时间错乱的量,然后将网络培训,同时在非喷射和喷射制度之间进行顺利的相互交错。我们把这一技术应用到传奇记忆股(LMUMU),以获得一个已知的混合SNNU优于S的常规结构的首例,包括LSTM、GRU和NRU,主要由于只有一小点精确度。我们从信号处理器处理信号处理工具,将神经活动功能激活功能功能功能功能作为临界点,然后在1比位上将活动降低1比重。