Quantization techniques applied to the inference of deep neural networks have enabled fast and efficient execution on resource-constraint devices. The success of quantization during inference has motivated the academic community to explore fully quantized training, i.e. quantizing back-propagation as well. However, effective gradient quantization is still an open problem. Gradients are unbounded and their distribution changes significantly during training, which leads to the need for dynamic quantization. As we show, dynamic quantization can lead to significant memory overhead and additional data traffic slowing down training. We propose a simple alternative to dynamic quantization, in-hindsight range estimation, that uses the quantization ranges estimated on previous iterations to quantize the present. Our approach enables fast static quantization of gradients and activations while requiring only minimal hardware support from the neural network accelerator to keep track of output statistics in an online fashion. It is intended as a drop-in replacement for estimating quantization ranges and can be used in conjunction with other advances in quantized training. We compare our method to existing methods for range estimation from the quantized training literature and demonstrate its effectiveness with a range of architectures, including MobileNetV2, on image classification benchmarks (Tiny ImageNet & ImageNet).
翻译:用于深神经网络推断的量化技术使资源约束装置能够快速高效地快速高效地执行资源约束装置。在推断过程中的量化成功促使学术界探索充分量化的培训,即对背面剖析。然而,有效的梯度量化仍然是一个尚未解决的问题。渐变量化方法没有限制,其分布变化很大,导致需要动态量化。正如我们所显示的那样,动态量化可导致大量存储管理以及额外的数据流量减缓培训。我们提出了动态量化的简单替代方法,即动态量化,即近视范围估算,即利用先前反复估算估计的量化范围对当前进行量化。我们的方法使得梯度和激活快速固定量化,同时只需要神经网络加速器提供最低限度的硬件支持,以在线跟踪产出统计数据。我们的目的是为了在量化范围上降低存储存储率,并可以与其他进展同步使用。我们的方法,包括图像网络模型模型的分类范围,我们将其现有方法与图象化模型的模型进行对比,以图象化分析范围为模型。