Neural network pruning is an essential approach for reducing the computational complexity of deep models so that they can be well deployed on resource-limited devices. Compared with conventional methods, the recently developed dynamic pruning methods determine redundant filters variant to each input instance which achieves higher acceleration. Most of the existing methods discover effective sub-networks for each instance independently and do not utilize the relationship between different inputs. To maximally excavate redundancy in the given network architecture, this paper proposes a new paradigm that dynamically removes redundant filters by embedding the manifold information of all instances into the space of pruned networks (dubbed as ManiDP). We first investigate the recognition complexity and feature similarity between images in the training set. Then, the manifold relationship between instances and the pruned sub-networks will be aligned in the training procedure. The effectiveness of the proposed method is verified on several benchmarks, which shows better performance in terms of both accuracy and computational cost compared to the state-of-the-art methods. For example, our method can reduce 55.3% FLOPs of ResNet-34 with only 0.57% top-1 accuracy degradation on ImageNet.
翻译:神经网络运行是降低深层模型的计算复杂性, 以便能在资源有限的装置上很好地部署。 与常规方法相比, 最近开发的动态运行方法确定每个输入实例的冗余过滤器变异, 从而实现更高的加速度。 大多数现有方法都为每个输入实例独立发现有效的子网络, 不使用不同输入之间的关系。 为了在特定网络结构中最大限度地挖掘冗余, 本文提出了一个新的模式, 通过将所有实例的多重信息嵌入纯净网络空间( 以 ManiDP 为底盘) 来动态地去除多余的过滤器。 我们首先调查了在培训集中图像的识别复杂性和特征。 然后, 实例和纯化子网络之间的多重关系将在培训程序中保持一致。 拟议方法的有效性在几个基准上得到验证, 这表明在准确性和计算成本方面, 与最新方法相比, 我们的方法可以减少ResNet-34 的55.3% FLOPs, 只有0.57 % 图像网络上最高-1 的图像降解率。