Structured pruning is an effective compression technique to reduce the computation of neural networks, which is usually achieved by adding perturbations to reduce network parameters at the cost of slightly increasing training loss. A more reasonable approach is to find a sparse minimizer along the flat minimum valley found by optimizers, i.e. stochastic gradient descent, which keeps the training loss constant. To achieve this goal, we propose the structured directional pruning based on orthogonal projecting the perturbations onto the flat minimum valley. We also propose a fast solver sDprun and further prove that it achieves directional pruning asymptotically after sufficient training. Experiments using VGG-Net and ResNet on CIFAR-10 and CIFAR-100 datasets show that our method obtains the state-of-the-art pruned accuracy (i.e. 93.97% on VGG16, CIFAR-10 task) without retraining. Experiments using DNN, VGG-Net and WRN28X10 on MNIST, CIFAR-10 and CIFAR-100 datasets demonstrate our method performs structured directional pruning, reaching the same minimum valley as the optimizer.
翻译:为实现这一目标,我们提议根据对平面最低谷地的扰动投影进行结构化的定向剪切,我们提议采用快速解析器SDprun,并进一步证明在经过充分培训后,它能达到方向性断线。 在CIFAR-10和CIFAR-100数据集上使用VGG-Net和ResNet的实验显示,我们的方法获得了最先进的截断精确度(即在VGGG16、CIFAR-10任务上获得93.97%的精确度),无需再培训。在MNIST、CIFAR-10和CIFAR-100数据库上使用DNNN、VGG-Net和WHRN28X1010进行实验,以最优化的方式运行。