Tensor decomposition is one of the fundamental technique for model compression of deep convolution neural networks owing to its ability to reveal the latent relations among complex structures. However, most existing methods compress the networks layer by layer, which cannot provide a satisfactory solution to achieve global optimization. In this paper, we proposed a model reduction method to compress the pre-trained networks using low-rank tensor decomposition of the convolution layers. Our method is based on the optimization techniques to select the proper ranks of decomposed network layers. A new regularization method, called funnel function, is proposed to suppress the unimportant factors during the compression, so the proper ranks can be revealed much easier. The experimental results show that our algorithm can reduce more model parameters than other tensor compression methods. For ResNet18 with ImageNet2012, our reduced model can reach more than twi times speed up in terms of GMAC with merely 0.7% Top-1 accuracy drop, which outperforms most existing methods in both metrics.
翻译:电离分解是深卷神经网络模型压缩的基本技术之一,因为它能够揭示复杂结构之间的潜在关系。 然而,大多数现有方法将网络层按层压缩,这无法提供实现全球优化的满意解决方案。 在本文中,我们提出了一个模型减少方法,用低调的电压分解组合层压缩预培训网络。我们的方法基于最优化技术,以选择分解网络层的适当级别。 一种叫做漏斗功能的新的正规化方法,在压缩过程中抑制不重要的因素,这样适当的排层可以更容易地披露。 实验结果显示,我们的算法可以比其他高压压方法减少更多的模型参数。 对于使用图像Net2012年的ResNet18,我们缩小的模型可以以仅0.7 % Top-1 的精度下降速度超过GMAC的两倍。