Currently, an increasing number of model pruning methods are proposed to resolve the contradictions between the computer powers required by the deep learning models and the resource-constrained devices. However, most of the traditional rule-based network pruning methods can not reach a sufficient compression ratio with low accuracy loss and are time-consuming as well as laborious. In this paper, we propose Automatic Block-wise and Channel-wise Network Pruning (ABCP) to jointly search the block-wise and channel-wise pruning action with deep reinforcement learning. A joint sample algorithm is proposed to simultaneously generate the pruning choice of each residual block and the channel pruning ratio of each convolutional layer from the discrete and continuous search space respectively. The best pruning action taking both the accuracy and the complexity of the model into account is obtained finally. Compared with the traditional rule-based pruning method, this pipeline saves human labor and achieves a higher compression ratio with lower accuracy loss. Tested on the mobile robot detection dataset, the pruned YOLOv3 model saves 99.5% FLOPs, reduces 99.5% parameters, and achieves 37.3 times speed up with only 2.8% mAP loss. The results of the transfer task on the sim2real detection dataset also show that our pruned model has much better robustness performance.
翻译:目前,提议了越来越多的模型裁剪方法,以解决深层学习模型和资源限制装置所要求的计算机能力之间的矛盾。 但是,大多数传统的基于规则的网络裁剪方法不能达到低精度损失的足够压缩率,而且耗时费力。 在本文中,我们提议采用自动阻截和频道操作网络预留法(ABCP),通过深层加固学习,共同搜索区块和频道操作操作动作。建议采用联合样本算法,同时从离散和连续搜索空间生成每个残余区块的裁剪切选择和每个相接层的频道裁剪切率。最终将获得考虑到模型精度和复杂性的最佳裁剪切行动。与传统的基于规则的裁剪裁方法相比,这一管道可以节省人类劳动力,并在降低精度损失的情况下实现更高的压缩率。在移动机器人探测数据集上测试的YOLOV3模型只节省99.5% FOPs,降低每个相隔断层层的频道裁剪切率率参数99.5 %; 也实现了将模型传输速度提高到2.8%的运行速度。