In this paper, we compress convolutional neural network (CNN) weights post-training via transform quantization. Previous CNN quantization techniques tend to ignore the joint statistics of weights and activations, producing sub-optimal CNN performance at a given quantization bit-rate, or consider their joint statistics during training only and do not facilitate efficient compression of already trained CNN models. We optimally transform (decorrelate) and quantize the weights post-training using a rate-distortion framework to improve compression at any given quantization bit-rate. Transform quantization unifies quantization and dimensionality reduction (decorrelation) techniques in a single framework to facilitate low bit-rate compression of CNNs and efficient inference in the transform domain. We first introduce a theory of rate and distortion for CNN quantization, and pose optimum quantization as a rate-distortion optimization problem. We then show that this problem can be solved using optimal bit-depth allocation following decorrelation by the optimal End-to-end Learned Transform (ELT) we derive in this paper. Experiments demonstrate that transform quantization advances the state of the art in CNN compression in both retrained and non-retrained quantization scenarios. In particular, we find that transform quantization with retraining is able to compress CNN models such as AlexNet, ResNet and DenseNet to very low bit-rates (1-2 bits).
翻译:在本文中,我们通过变换量化来压缩进化神经网络(CNN)重量。 先前的CNN量化技术往往忽视重量和激活的联合统计,在给定的四分制位速率中产生亚最佳CNN性表现,或者仅仅在训练期间考虑其联合统计数据,不利于有效压缩已经受过训练的CNN模型。我们优化地转换( 降压) 和量化后训练重量。我们使用一个比率扭曲框架来改进任何特定量化位数化位数的压缩。 变换四分化统一量化和元化减少( 变换) 网络化技术, 在一个单一框架内, 以方便低位位调压缩CNNs和在变异域中有效推断, 或只考虑其联合统计数据, 并且不便于有效压缩已经受过训练的CNNNCN的模型。 我们随后展示了这一问题可以通过最佳的州级到级的低级变换版( ELT) 和 网络化( 网络变版) 技术在本文中以最优化的变版变版性变版式变版模式来解决该问题。