论文标题
在神经形态硬件上的机器学习工作负载的节能部署
Energy-Efficient Deployment of Machine Learning Workloads on Neuromorphic Hardware
论文作者
论文摘要
随着技术行业正在朝着实施诸如较小边缘计算设备的自然语言处理,路径计划,图像分类等任务迈进,对算法和硬件加速器的更有效实施的需求已成为重要的研究领域。近年来,已经发布了几种Edge深入学习硬件加速器,这些硬件加速器特别着重于降低深神经网络(DNNS)所消耗的功能和区域。另一方面,在离散的时间序列数据上运行的尖峰神经网络(SNN)已被证明可以在基于专门的神经形态事件/基于异常的硬件上部署时,即使是上述边缘DNN加速器也可以实现大量功率降低。尽管神经形态硬件表现出了在边缘加速深度学习任务的巨大潜力,但算法和硬件的当前空间是有限的,并且仍处于早期开发状态。因此,已经提出了许多旨在将预训练的DNN转换为SNN的混合方法。在这项工作中,我们提供了将预训练的DNN转换为SNN的一般指南,同时还提出了提高有关延迟,功率和能量的神经形态硬件的转换SNN的部署的技术。我们的实验结果表明,与英特尔神经计算棒2相比,英特尔的神经形态处理器Loihi使用我们的SNN改进技术,在测试的图像分类任务中消耗的功率降低了27倍,减少了5倍。
As the technology industry is moving towards implementing tasks such as natural language processing, path planning, image classification, and more on smaller edge computing devices, the demand for more efficient implementations of algorithms and hardware accelerators has become a significant area of research. In recent years, several edge deep learning hardware accelerators have been released that specifically focus on reducing the power and area consumed by deep neural networks (DNNs). On the other hand, spiking neural networks (SNNs) which operate on discrete time-series data, have been shown to achieve substantial power reductions over even the aforementioned edge DNN accelerators when deployed on specialized neuromorphic event-based/asynchronous hardware. While neuromorphic hardware has demonstrated great potential for accelerating deep learning tasks at the edge, the current space of algorithms and hardware is limited and still in rather early development. Thus, many hybrid approaches have been proposed which aim to convert pre-trained DNNs into SNNs. In this work, we provide a general guide to converting pre-trained DNNs into SNNs while also presenting techniques to improve the deployment of converted SNNs on neuromorphic hardware with respect to latency, power, and energy. Our experimental results show that when compared against the Intel Neural Compute Stick 2, Intel's neuromorphic processor, Loihi, consumes up to 27x less power and 5x less energy in the tested image classification tasks by using our SNN improvement techniques.