The success of convolutional neural networks (CNNs) in various applications is accompanied by a significant increase in computation and parameter storage costs. Recent efforts to reduce these overheads involve pruning and compressing the weights of various layers while at the same time aiming to not sacrifice performance. In this paper, we propose a novel criterion for CNN pruning inspired by neural network interpretability: The most relevant units, i.e. weights or filters, are automatically found using their relevance scores obtained from concepts of explainable AI (XAI). By exploring this idea, we connect the lines of interpretability and model compression research. We show that our proposed method can efficiently prune CNN models in transfer-learning setups in which networks pre-trained on large corpora are adapted to specialized tasks. The method is evaluated on a broad range of computer vision datasets. Notably, our novel criterion is not only competitive or better compared to state-of-the-art pruning criteria when successive retraining is performed, but clearly outperforms these previous criteria in the resource-constrained application scenario in which the data of the task to be transferred to is very scarce and one chooses to refrain from fine-tuning. Our method is able to compress the model iteratively while maintaining or even improving accuracy. At the same time, it has a computational cost in the order of gradient computation and is comparatively simple to apply without the need for tuning hyperparameters for pruning.
翻译:在各种应用中的神经神经网络(CNNs)成功的同时,计算和参数存储成本也大幅增加。最近为减少这些管理费用而作的努力包括裁剪和压缩不同层的重量,同时力求不牺牲性能。在本文件中,我们提出了一个受神经网络解释性启发的CNN运行新标准:最相关的单位,即重量或过滤器,在进行连续再培训时,其相关性分数自动被发现,但是通过探索这一想法,我们将可解释性和模型压缩研究的线连接起来。我们表明,我们拟议的方法可以有效地在转移学习设置中采用CNN模型,在这种设置中,网络在大体形体上预先训练的网络可以适应专门的任务。该方法在广泛的计算机视觉数据集上得到评估:值得注意的是,在连续再培训时,我们的新标准不仅具有竞争力或更好,而且显然超越了在资源限制的应用方案中的这些以前标准。在这种情况下,我们所提出的方法可以有效使用CNN模型模型模型模型的模型模型模型模型,在不精确性调整过程中可以避免调整,而在不断调整的计算中选择一种方法,在不断调整时,在不断调整时,在不断调整的递增缩。