论文标题
性能激增:具有量化激活功能的训练混合兴奋的神经网络
A Spike in Performance: Training Hybrid-Spiking Neural Networks with Quantized Activation Functions
论文作者
论文摘要
机器学习社区已经对神经网络的能源效率越来越感兴趣。尖峰神经网络(SNN)是一种有前途的能源计算方法,因为将其激活水平量化为暂时稀疏的一位值(即“尖峰”事件),该值还将重量活性产品的总和转换为简单的重量(每个峰值的重量)。但是,将非加速网络转换为SNN时保持最先进的(SOTA)准确性的目的是难以捉摸的挑战,这主要是由于尖峰只有一点点精度。我们从信号处理中采用工具,将神经激活作为具有时间扩散误差的量化器,然后训练网络,同时在非尖峰和尖峰方面平稳插值。我们将此技术应用于Legendre存储单元(LMU),以获得混合SNN的第一个已知示例,其准确性优于SOTA复发架构(包括LSTM,GRU和NRU),而将活动降低到最多3.74位,平均为1.26位,每位重量为1.26位。我们讨论这些方法如何显着提高神经网络的能源效率。
The machine learning community has become increasingly interested in the energy efficiency of neural networks. The Spiking Neural Network (SNN) is a promising approach to energy-efficient computing, since its activation levels are quantized into temporally sparse, one-bit values (i.e., "spike" events), which additionally converts the sum over weight-activity products into a simple addition of weights (one weight for each spike). However, the goal of maintaining state-of-the-art (SotA) accuracy when converting a non-spiking network into an SNN has remained an elusive challenge, primarily due to spikes having only a single bit of precision. Adopting tools from signal processing, we cast neural activation functions as quantizers with temporally-diffused error, and then train networks while smoothly interpolating between the non-spiking and spiking regimes. We apply this technique to the Legendre Memory Unit (LMU) to obtain the first known example of a hybrid SNN outperforming SotA recurrent architectures -- including the LSTM, GRU, and NRU -- in accuracy, while reducing activities to at most 3.74 bits on average with 1.26 significant bits multiplying each weight. We discuss how these methods can significantly improve the energy efficiency of neural networks.