Low-bit quantization of network weights and activations can drastically reduce the memory footprint, complexity, energy consumption and latency of Deep Neural Networks (DNNs). However, low-bit quantization can also cause a considerable drop in accuracy, in particular when we apply it to complex learning tasks or lightweight DNN architectures. In this paper, we propose a training procedure that relaxes the low-bit quantization. We call this procedure \textit{DNN Quantization with Attention} (DQA). The relaxation is achieved by using a learnable linear combination of high, medium and low-bit quantizations. Our learning procedure converges step by step to a low-bit quantization using an attention mechanism with temperature scheduling. In experiments, our approach outperforms other low-bit quantization techniques on various object recognition benchmarks such as CIFAR10, CIFAR100 and ImageNet ILSVRC 2012, achieves almost the same accuracy as a full precision DNN, and considerably reduces the accuracy drop when quantizing lightweight DNN architectures.
翻译:网络重量和激活的低位量化可以大幅降低深神经网络(DNN)的记忆足迹、复杂性、能量消耗和延迟度。然而,低位量化也可以导致精确度大幅下降,特别是当我们将其应用到复杂的学习任务或轻量的 DNN 结构时。在本文中,我们建议了一个能放松低位量化的培训程序。我们称这个程序为\ textit{DNN 量度化并注意(DQA ) 。通过使用高、中、低位四分制的可学习线性组合来实现放松。我们的学习程序会一步地集中到低位量化,使用温度表的注意机制。在实验中,我们的方法在各种物体识别基准上比其他低位量化技术(如CIFAR10, CIFAR100 和图像网 ILSVRC 2012 ) 取得了与完全精确的 DNN(DN)几乎相同的精确度,并在对轻量 DNN 结构进行四分时大大降低了精度下降的精度。