This paper proposes a training method having multiple cyclic training for achieving enhanced performance in low-bit quantized convolutional neural networks (CNNs). Quantization is a popular method for obtaining lightweight CNNs, where the initialization with a pretrained model is widely used to overcome degraded performance in low-resolution quantization. However, large quantization errors between real values and their low-bit quantized ones cause difficulties in achieving acceptable performance for complex networks and large datasets. The proposed training method softly delivers the knowledge of pretrained models to low-bit quantized models in multiple quantization steps. In each quantization step, the trained weights of a model are used to initialize the weights of the next model with the quantization bit depth reduced by one. With small change of the quantization bit depth, the performance gap can be bridged, thus providing better weight initialization. In cyclic training, after training a low-bit quantized model, its trained weights are used in the initialization of its accurate model to be trained. By using better training ability of the accurate model in an iterative manner, the proposed method can produce enhanced trained weights for the low-bit quantized model in each cycle. Notably, the training method can advance Top-1 and Top-5 accuracies of the binarized ResNet-18 on the ImageNet dataset by 5.80% and 6.85%, respectively.
翻译:本文提出一种具有多种周期性培训的培训方法,以在低位四分级的进化神经网络(CNNs)中提高性能。量化是一种获得轻量级CNN的流行方法,在低分辨率四分制中,先行模式的初始化被广泛用于克服退化的性能。然而,实际值与其低位四分制型之间的重大量化错误,给复杂网络和大型数据集达到可接受的性能带来困难。拟议的培训方法以软方式向多个四分制步骤的低位网络化模型提供预先培训模型的知识。在每一个四分制步骤中,都使用经过培训的模型的重量来初始化下一个模型的重量,同时将四分化的深度降低一个。随着四分制深度的微小变化,性能差距可以弥合,从而提供更好的权重度初始化。在培训低位数分级模型的初始化过程中,将经过培训的重度用于其精确模型的初始化。在每一个四分位化模型的初始化阶段使用经过培训能力,通过经过更精确的五分位化的高级培训的模型,以更高比例的方式制作一个高级版本的模型。