Pruning techniques are used comprehensively to compress convolutional neural networks (CNNs) on image classification. However, the majority of pruning methods require a well pre-trained model to provide useful supporting parameters, such as C1-norm, BatchNorm value and gradient information, which may lead to inconsistency of filter evaluation if the parameters of the pre-trained model are not well optimized. Therefore, we propose a sensitiveness based method to evaluate the importance of each layer from the perspective of inference accuracy by adding extra damage for the original model. Because the performance of the accuracy is determined by the distribution of parameters across all layers rather than individual parameter, the sensitiveness based method will be robust to update of parameters. Namely, we can obtain similar importance evaluation of each convolutional layer between the imperfect-trained and fully trained models. For VGG-16 on CIFAR-10, even when the original model is only trained with 50 epochs, we can get same evaluation of layer importance as the results when the model is trained fully. Then we will remove filters proportional from each layer by the quantified sensitiveness. Our sensitiveness based pruning framework is verified efficiently on VGG-16, a customized Conv-4 and ResNet-18 with CIFAR-10, MNIST and CIFAR-100, respectively.
翻译:在图像分类方面,对压缩进化神经网络(CNNs)全面使用节制技术,压缩进化神经网络(CNNs)的图像分类。然而,大多数修剪方法都需要经过良好预先培训的模式,以提供有用的支持参数,如C1-诺尔姆、批量Norm值和梯度信息,如果未对预培训模型的参数进行最佳优化,可能会导致过滤评价不一致。因此,我们建议一种基于敏感的方法,从推断准确性的角度来评估每一层的重要性,方法是为原始模型增加额外的损害。由于准确性的表现取决于所有层的参数分布,而不是单个参数的分布,基于敏感度的基础方法将有力更新参数。也就是说,我们可以对不完善和受过充分训练的模型之间的每个进化层进行类似的重要评价。对于CIFAR-10的VGG-16,即使最初模型仅经过50个小节点的培训,我们也能得到与模型充分培训后的结果一样的层次重要性评价。然后,我们将通过量化的敏感度,从每个层中去除的过滤器,我们基于精度的敏感度和REAR-16分别与IM-CR的S-CRAR的精度框架进行有效的核查。