In this paper, we propose an ultrafast automated model compression framework called SeerNet for flexible network deployment. Conventional non-differen-tiable methods discretely search the desirable compression policy based on the accuracy from exhaustively trained lightweight models, and existing differentiable methods optimize an extremely large supernet to obtain the required compressed model for deployment. They both cause heavy computational cost due to the complex compression policy search and evaluation process. On the contrary, we obtain the optimal efficient networks by directly optimizing the compression policy with an accurate performance predictor, where the ultrafast automated model compression for various computational cost constraint is achieved without complex compression policy search and evaluation. Specifically, we first train the performance predictor based on the accuracy from uncertain compression policies actively selected by efficient evolutionary search, so that informative supervision is provided to learn the accurate performance predictor with acceptable cost. Then we leverage the gradient that maximizes the predicted performance under the barrier complexity constraint for ultrafast acquisition of the desirable compression policy, where adaptive update stepsizes with momentum are employed to enhance optimality of the acquired pruning and quantization strategy. Compared with the state-of-the-art automated model compression methods, experimental results on image classification and object detection show that our method achieves competitive accuracy-complexity trade-offs with significant reduction of the search cost.
翻译:在本文中,我们提出了一个名为SeerNet的超快速自动化模型压缩框架,以进行网络部署的灵活性。传统的非可微方法根据从穷举训练的轻量级模型中获得的准确性,离散地搜索可取的压缩策略,而现有的可微方法则优化一个极为庞大的超网络,以获得所需的压缩模型进行部署。这两种方法都会因复杂的压缩策略搜索和评估过程而导致重大的计算成本。相反,我们通过直接优化具有准确性能预测器的压缩策略,从而获得优化高效网络所需的最佳方式,从而实现了各种计算成本约束下的超快速自动化模型压缩,而无需进行复杂的压缩策略搜索和评估。具体来说,我们首先在使用高效的进化搜索主动选择的不确定压缩策略的准确性的基础上训练性能预测器,从而提供了信息丰富的监督来学习具有可接受成本的准确性能预测器。然后,我们利用梯度来最大化在障碍复杂度约束下的预测性能,以超快速获取所需的压缩策略,并采用动量自适应更新步幅来增强所获得的剪枝和量化策略的最优性。与最先进的自动模型压缩方法相比,分类和目标检测的实验结果表明,我们的方法在显著降低搜索成本的同时实现了具有竞争力的准确性-复杂性平衡。