Recent studies on deep convolutional neural networks present a simple paradigm of architecture design, i.e., models with more MACs typically achieve better accuracy, such as EfficientNet and RegNet. These works try to enlarge all the stages in the model with one unified rule by sampling and statistical methods. However, we observe that some network architectures have similar MACs and accuracies, but their allocations on computations for different stages are quite different. In this paper, we propose to enlarge the capacity of CNN models by improving their width, depth and resolution on stage level. Under the assumption that the top-performing smaller CNNs are a proper subcomponent of the top-performing larger CNNs, we propose an greedy network enlarging method based on the reallocation of computations. With step-by-step modifying the computations on different stages, the enlarged network will be equipped with optimal allocation and utilization of MACs. On EfficientNet, our method consistently outperforms the performance of the original scaling method. In particular, with application of our method on GhostNet, we achieve state-of-the-art 80.9% and 84.3% ImageNet top-1 accuracies under the setting of 600M and 4.4B MACs, respectively.
翻译:最近对深层进化神经网络的研究提出了建筑设计的一个简单范例,即拥有更多MAC的模型通常会达到更高的准确度,例如高效网络和RegNet。这些工程试图通过抽样和统计方法扩大模型的所有阶段,通过统一规则通过抽样和统计方法扩大模型的所有阶段。然而,我们注意到,有些网络结构具有类似的MAC和理解,但它们在不同阶段的计算分配情况却大不相同。在本文件中,我们提议通过改进CNN模型的广度、深度和在阶段一级的分辨率来扩大其能力。根据最优秀的小型CNN模型是最优秀的CNN系统的适当子组成部分这一假设,我们提议以计算再分配为基础扩大贪婪的网络方法。随着对不同阶段的计算进行逐步修改,扩大的网络将配备最优化的MACs分配和利用。关于高效网络,我们的方法始终比最初的缩放方法的绩效高。特别是在GhindNet上应用我们的方法,我们实现了第80.9%和84.3%的图像网络在4.4%和4.4%的图像网络上分别设定了第4.4A和4.4%的状态。