Deep convolutional neural networks are shown to be overkill with high parametric and computational redundancy in many application scenarios, and an increasing number of works have explored model pruning to obtain lightweight and efficient networks. However, most existing pruning approaches are driven by empirical heuristics and rarely consider the joint impact of channels, leading to unguaranteed and suboptimal performance. In this paper, we propose a novel channel pruning method via class-aware trace ratio optimization (CATRO) to reduce the computational burden and accelerate the model inference. Utilizing class information from a few samples, CATRO measures the joint impact of multiple channels by feature space discriminations and consolidates the layer-wise impact of preserved channels. By formulating channel pruning as a submodular set function maximization problem, CATRO solves it efficiently via a two-stage greedy iterative optimization procedure. More importantly, we present theoretical justifications on convergence and performance of CATRO. Experimental results demonstrate that CATRO achieves higher accuracy with similar computation cost or lower computation cost with similar accuracy than other state-of-the-art channel pruning algorithms. In addition, because of its class-aware property, CATRO is suitable to prune efficient networks adaptively for various classification subtasks, enhancing handy deployment and usage of deep networks in real-world applications.
翻译:深相神经网络在很多应用情景中被证明是超权的,具有高参数和计算冗余,而且越来越多的工程已经探索了模型裁剪,以获得轻量和高效的网络;然而,大多数现有裁剪方法都是由经验超常驱动的,很少考虑频道的联合影响,导致共和和和不优化的性能。在本文中,我们提议了一种新型的频道裁剪方法,即通过级觉微量比优化(CATRO)来减少计算负担并加速模型推断。利用少数样本的类信息,CATRO通过地貌空间歧视来测量多个频道的联合影响,并巩固保留频道的层性影响。通过将频道裁剪裁作为子组合功能的最大化问题,CATRO通过两个阶段贪婪迭代优化程序有效地解决了这一问题。更重要的是,我们为CATRO的趋同和业绩优化提供了理论依据。实验结果表明,CATRO的计算成本或计算成本较低,与其他州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-级-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州-州