With the surging popularity of edge computing, the need to efficiently perform neural network inference on battery-constrained IoT devices has greatly increased. While algorithmic developments enable neural networks to solve increasingly more complex tasks, the deployment of these networks on edge devices can be problematic due to the stringent energy, latency, and memory requirements. One way to alleviate these requirements is by heavily quantizing the neural network, i.e. lowering the precision of the operands. By taking quantization to the extreme, e.g. by using binary values, new opportunities arise to increase the energy efficiency. Several hardware accelerators exploiting the opportunities of low-precision inference have been created, all aiming at enabling neural network inference at the edge. In this chapter, design choices and their implications on the flexibility and energy efficiency of several accelerators supporting extremely quantized networks are reviewed.
翻译:随着边缘计算越来越受欢迎,对受电池限制的IoT装置有效进行神经网络推断的必要性大大增加。虽然算法的发展使神经网络能够解决日益复杂的任务,但由于严格的能量、延缓性和记忆要求,在边缘装置上部署这些网络可能会有问题。缓解这些要求的方法之一是对神经网络进行大量量化,即降低操作的精确度。通过将量化到极端,例如使用二元值,出现了提高能源效率的新机会。已经创造了几个利用低精度推断机会的硬件加速器,其目的都是在边缘使神经网络能够进行推断。在本章中,审查了设计的选择及其对支持极小四分化网络的若干加速器的灵活性和能源效率的影响。