Structured channel pruning has been shown to significantly accelerate inference time for convolution neural networks (CNNs) on modern hardware, with a relatively minor loss of network accuracy. Recent works permanently zero these channels during training, which we observe to significantly hamper final accuracy, particularly as the fraction of the network being pruned increases. We propose Soft Masking for cost-constrained Channel Pruning (SMCP) to allow pruned channels to adaptively return to the network while simultaneously pruning towards a target cost constraint. By adding a soft mask re-parameterization of the weights and channel pruning from the perspective of removing input channels, we allow gradient updates to previously pruned channels and the opportunity for the channels to later return to the network. We then formulate input channel pruning as a global resource allocation problem. Our method outperforms prior works on both the ImageNet classification and PASCAL VOC detection datasets.
翻译:结构化的频道修剪显示,在现代硬件上大幅加快神经网络卷发的推算时间,相对而言网络准确性损失较小。最近,在培训期间,这些频道一直处于永久零状态,我们观察到,这严重妨碍了最终准确性,特别是由于网络的一小部分被修补。我们提议为成本限制的频道预留(SMCP)进行软遮罩,以便经修补的频道能够适应性地返回网络,同时运行到目标成本限制。通过从删除输入渠道的角度对重量进行软面罩重新校准和频道修补,我们允许对以前经修补的频道进行梯度更新,并让频道有机会以后返回网络。我们随后将输入通道的修剪成一个全球资源分配问题。我们的方法超越了以前在图像网络分类和PASAL VOC探测数据集上的工作。