We propose a novel, structured pruning algorithm for neural networks -- the iterative, Sparse Structured Pruning algorithm, dubbed as i-SpaSP. Inspired by ideas from sparse signal recovery, i-SpaSP operates by iteratively identifying a larger set of important parameter groups (e.g., filters or neurons) within a network that contribute most to the residual between pruned and dense network output, then thresholding these groups based on a smaller, pre-defined pruning ratio. For both two-layer and multi-layer network architectures with ReLU activations, we show the error induced by pruning with i-SpaSP decays polynomially, where the degree of this polynomial becomes arbitrarily large based on the sparsity of the dense network's hidden representations. In our experiments, i-SpaSP is evaluated across a variety of datasets (i.e., MNIST and ImageNet) and architectures (i.e., feed forward networks, ResNet34, and MobileNetV2), where it is shown to discover high-performing sub-networks and improve upon the pruning efficiency of provable baseline methodologies by several orders of magnitude. Put simply, i-SpaSP is easy to implement with automatic differentiation, achieves strong empirical results, comes with theoretical convergence guarantees, and is efficient, thus distinguishing itself as one of the few computationally efficient, practical, and provable pruning algorithms.
翻译:我们为神经网络提议了一个创新的、结构化的神经网络运行算法 -- -- 迭代的、Sprassy结构化的和多层的网络结构算法,称为i-SpaSP。在信号恢复稀少的理念的启发下,i-SpaSP在网络内运作,迭接地确定一系列更重要的参数组(例如过滤器或神经元),在网络中最有助于处理网络输出和密集输出之间的剩余部分,然后根据一个更小的、预先界定的裁剪率比率来对这些组进行阈值。对于使用ReLU激活的双层和多层网络结构结构,我们展示了由i-SpaSP运行的运行导致的误差,在多层信号恢复后,根据密度网络隐藏的表达方式的宽广性,这种多层多层模型的高度变得任意性大。 在我们的实验中, i-SpaSP在一系列数据集(即MNIST和图像网络)和结构(即向前端网络、ResNet34和移动网络2)的启动者,我们展示了这种错误的运行的运行过程效率,通过简单化的精确的精确的精确的精确排序,从而可以实现高水平的精确的精确的精确的精确排序,从而实现某种的精确的精确的精确的精确的精确的精确的计算方法和精确的精确的精确的精确的精确的计算结果。