Model parameter regularization is a widely used technique to improve generalization, but also can be used to shape the weight distributions for various purposes. In this work, we shed light on how weight regularization can assist model quantization and compression techniques, and then propose range regularization (R^2) to further boost the quality of model optimization by focusing on the outlier prevention. By effectively regulating the minimum and maximum weight values from a distribution, we mold the overall distribution into a tight shape so that model compression and quantization techniques can better utilize their limited numeric representation powers. We introduce L-inf regularization, its extension margin regularization and a new soft-min-max regularization to be used as a regularization loss during full-precision model training. Coupled with state-of-the-art quantization and compression techniques, models trained with R^2 perform better on an average, specifically at lower bit weights with 16x compression ratio. We also demonstrate that R^2 helps parameter constrained models like MobileNetV1 achieve significant improvement of around 8% for 2 bit quantization and 7% for 1 bit compression.
翻译:模型参数的正规化是一种广泛应用的技术,可以改进一般化,但也可以用来为各种目的塑造重量分布。在这项工作中,我们说明了重量正规化如何有助于模型量化和压缩技术,然后建议范围正规化(R2/2),以便通过注重外部预防进一步提高模型优化的质量。我们通过有效地调节分布中最小和最大重量值,将整体分布模拟成一个紧凑的形状,以便模型压缩和定量化技术能够更好地利用其有限的数字代表能力。我们引入了L-inf正规化、其扩展边距正规化和新的软最小值正规化,以便在全面精密模型培训中用作正规化损失。与最先进的量化和压缩技术相结合,经过R%2培训的模型在平均上表现更好,特别是在低位重量和16x压缩比率下。我们还证明R%2帮助诸如PlopNetV1号的受限参数模型在2位化和1位压缩中取得了约8%的显著改进。</s>