Recently, tile pruning has been widely studied to accelerate the inference of deep neural networks (DNNs). However, we found that the loss due to tile pruning, which can eliminate important elements together with unimportant elements, is large on trained DNNs. In this study, we propose a one-shot reparameterization method, called TileTrans, to reduce the loss of tile pruning. Specifically, we repermute the rows or columns of the weight matrix such that the model architecture can be kept unchanged after reparameterization. This repermutation realizes the reparameterization of the DNN model without any retraining. The proposed reparameterization method combines important elements into the same tile; thus, preserving the important elements after the tile pruning. Furthermore, TileTrans can be seamlessly integrated into existing tile pruning methods because it is a pre-processing method executed before pruning, which is orthogonal to most existing methods. The experimental results demonstrate that our method is essential in reducing the loss of tile pruning on DNNs. Specifically, the accuracy is improved by up to 17% for AlexNet while 5% for ResNet-34, where both models are pre-trained on ImageNet.
翻译:最近,人们广泛研究了瓷砖的修剪方法,以加速深神经网络(DNNs)的推导。然而,我们发现,由于瓷砖修剪而导致的可消除重要元素和不重要元素的重新修剪损失在经过训练的DNNs上是巨大的。在这个研究中,我们提议了一个叫做TileTrans的一发再分计法,以降低瓷砖剪裁机的损失。具体地说,我们重新对重力矩阵的行或列进行重新检查,这样模型结构在重新校准后可以保持不变。这种重新修整可以实现DNN模型的再修补。拟议的重新修补方法将重要元素与不重要元素合并到同一瓷砖剪裁后的重要元素。此外,TileTrans可以顺利地融入现有的瓷砖剪裁方法,因为它是预处理方法,在重新校正后,模型结构结构可以保持不变。这种方法对于减少DNNM模型的损失至关重要,而亚历克斯特的精确度则是在5Net上进行精确度测试。