Neural network pruning is frequently used to compress over-parameterized networks by large amounts, while incurring only marginal drops in generalization performance. However, the impact of pruning on networks that have been highly optimized for efficient inference has not received the same level of attention. In this paper, we analyze the effect of pruning for computer vision, and study state-of-the-art ConvNets, such as the FBNetV3 family of models. We show that model pruning approaches can be used to further optimize networks trained through NAS (Neural Architecture Search). The resulting family of pruned models can consistently obtain better performance than existing FBNetV3 models at the same level of computation, and thus provide state-of-the-art results when trading off between computational complexity and generalization performance on the ImageNet benchmark. In addition to better generalization performance, we also demonstrate that when limited computation resources are available, pruning FBNetV3 models incur only a fraction of GPU-hours involved in running a full-scale NAS.
翻译:神经网络修剪常常被用来大量压缩超分度网络,但一般性能只产生边际下降。然而,对为高效推断而高度优化的网络的修剪影响并没有得到同等程度的注意。在本文中,我们分析了计算机视觉修剪的效果,并研究了先进的ConvNets,例如FBNetV3模型系列。我们表明,模型修剪方法可用于进一步优化通过NAS(神经建筑搜索)培训的网络。因此,在相同的计算水平上,经修剪的模型的组合可以始终比现有的FBNetV3模型取得更好的性能,因此,在将图像网络基准的计算复杂性和一般性能进行交换时,提供了最先进的结果。除了提高通用性能外,我们还证明,在有限的计算资源可用的情况下,运行FBNetV3模型只能产生部分的GPU小时,用于全面运行NAS。