The existence of a lot of redundant information in convolutional neural networks leads to the slow deployment of its equipment on the edge. To solve this issue, we proposed a novel deep learning model compression acceleration method based on data distribution characteristics, namely Pruning Filter via Gaussian Distribution Feature(PFGDF) which was to found the smaller interval of the convolution layer of a certain layer to describe the original on the grounds of distribution characteristics . Compared with revious advanced methods, PFGDF compressed the model by filters with insignificance in distribution regardless of the contribution and sensitivity information of the convolution filter. The pruning process of the model was automated, and always ensured that the compressed model could restore the performance of original model. Notably, on CIFAR-10, PFGDF compressed the convolution filter on VGG-16 by 66:62%, the parameter reducing more than 90%, and FLOPs achieved 70:27%. On ResNet-32, PFGDF reduced the convolution filter by 21:92%. The parameter was reduced to 54:64%, and the FLOPs exceeded 42%
翻译:在卷发神经网络中存在大量冗余信息,导致其设备在边缘的部署速度缓慢。为了解决这一问题,我们提议了一个基于数据分配特点的新型深学习模型压缩加速法,即通过高森分布特性的普鲁宁过滤器(PFGDF),该方法将某个层的细卷变层间隔以分布特性为由描述原始数据。与废弃的先进方法相比,FFFGDF通过分配无重的过滤器压缩模型,而不管卷发过滤器的贡献和敏感性信息如何。该模型的运行过程是自动化的,并且始终确保压缩模型能够恢复原始模型的性能。值得注意的是,在CIFAR-10, PFGDF将VG-16上的卷变过滤器压缩了66:62%,参数减少了90%以上,FLOPs实现了70:27%。在ResNet-32上,PFGDF将模型的过滤器减少了21:92%。参数减少到54:64%,FLOPs超过了42%。