Pruning on neural networks before training not only compresses the original models, but also accelerates the network training phase, which has substantial application value. The current work focuses on fine-grained pruning, which uses metrics to calculate weight scores for weight screening, and extends from the initial single-order pruning to iterative pruning. Through these works, we argue that network pruning can be summarized as an expressive force transfer process of weights, where the reserved weights will take on the expressive force from the removed ones for the purpose of maintaining the performance of original networks. In order to achieve optimal expressive force scheduling, we propose a pruning scheme before training called Neural Network Panning which guides expressive force transfer through multi-index and multi-process steps, and designs a kind of panning agent based on reinforcement learning to automate processes. Experimental results show that Panning performs better than various available pruning before training methods.
翻译:在培训前对神经网络进行控制,不仅压缩原始模型,而且加速具有大量应用价值的网络培训阶段。目前的工作侧重于精细的裁剪,它使用量度计算重量分以进行体重筛选,从最初的单级裁剪到迭接剪裁,通过这些工程,我们可以将网络的修剪可归纳为重力的表达力转换过程,保留重量将从已删除的重力中取出,以维持原始网络的性能。为了实现最佳的表达力排期,我们提议在培训称为神经网络的抽动前先进行一个剪裁计划,以引导通过多指数和多过程步骤的表达力转移,并设计一种基于强化学习到自动过程的跨动剂。实验结果显示,潘宁的表现比培训方法之前的各种现有剪裁效果要好。