Neural Architecture Search (NAS) is quickly becoming the standard methodology to design neural network models. However, NAS is typically compute-intensive because multiple models need to be evaluated before choosing the best one. To reduce the computational power and time needed, a proxy task is often used for evaluating each model instead of full training. In this paper, we evaluate conventional reduced-training proxies and quantify how well they preserve ranking between multiple models during search when compared with the rankings produced by final trained accuracy. We propose a series of zero-cost proxies, based on recent pruning literature, that use just a single minibatch of training data to compute a model's score. Our zero-cost proxies use 3 orders of magnitude less computation but can match and even outperform conventional proxies. For example, Spearman's rank correlation coefficient between final validation accuracy and our best zero-cost proxy on NAS-Bench-201 is 0.82, compared to 0.61 for EcoNAS (a recently proposed reduced-training proxy). Finally, we use these zero-cost proxies to enhance existing NAS search algorithms such as random search, reinforcement learning, evolutionary search and predictor-based search. For all search methodologies and across three different NAS datasets, we are able to significantly improve sample efficiency, and thereby decrease computation, by using our zero-cost proxies. For example on NAS-Bench-101, we achieved the same accuracy 4$\times$ quicker than the best previous result. Our code is made public at: https://github.com/mohsaied/zero-cost-nas.
翻译:神经结构搜索(NAS)正在迅速成为设计神经网络模型的标准方法。 但是,NAS通常需要大量计算,因为需要对多个模型进行计算,然后才能选择最佳模型。为了减少计算力和时间,通常使用代理任务来评价每个模型,而不是全面培训。在本文中,我们评估常规的减少培训代理,并量化在搜索过程中,与最终经过培训的准确性排名相比,它们保持多种模型之间的排名有多好。我们根据最近的修剪文献,提议了一系列零成本代理数据,仅仅使用一小批培训数据来计算模型的分数。为了减少计算,我们的零成本代理数据需要使用3个数量级的计算,但可以匹配甚至超过常规代理。例如,Spearman的等级相关系数在最终验证准确性与我们在NAS-Bench-201上的最佳零成本代数之间是0.82,而EcoNAS(最近提出的降低培训代理)则为0.61。最后,我们使用这些零成本代理数据搜索工具来增强现有的NAS-101的计算效率,我们通过随机搜索和升级的方法来大幅改进了所有以进行搜索。