Modern deep neural networks are often too large to use in many practical scenarios. Neural network pruning is an important technique for reducing the size of such models and accelerating inference. Gibbs pruning is a novel framework for expressing and designing neural network pruning methods. Combining approaches from statistical physics and stochastic regularization methods, it can train and prune a network simultaneously in such a way that the learned weights and pruning mask are well-adapted for each other. It can be used for structured or unstructured pruning and we propose a number of specific methods for each. We compare our proposed methods to a number of contemporary neural network pruning methods and find that Gibbs pruning outperforms them. In particular, we achieve a new state-of-the-art result for pruning ResNet-56 with the CIFAR-10 dataset.
翻译:现代深神经网络往往太大,无法用于许多实际情景。神经网络运行是缩小此类模型规模和加速推断的重要技术。 Gibbs 运行是表达和设计神经网络运行方法的新框架。 将统计物理和随机调节方法相结合,它可以同时培训和利用一个网络,使学到的重量和修剪面罩相互适应,可以用于结构化或非结构化的剪裁,我们为每个网络提出若干具体的方法。 我们将我们建议的方法与当代神经网络运行方法进行比较,并发现Gibs的剪裁方法优于这些方法。 特别是,我们通过CIFAR-10数据集,实现了对ResNet-56进行剪接的新的最新结果。