Neural Architecture Search (NAS) is quickly becoming the standard methodology to design neural network models. However, NAS is typically compute-intensive because multiple models need to be evaluated before choosing the best one. To reduce the computational power and time needed, a proxy task is often used for evaluating each model instead of full training. In this paper, we evaluate conventional reduced-training proxies and quantify how well they preserve ranking between multiple models during search when compared with the rankings produced by final trained accuracy. We propose a series of zero-cost proxies, based on recent pruning literature, that use just a single minibatch of training data to compute a model's score. Our zero-cost proxies use 3 orders of magnitude less computation but can match and even outperform conventional proxies. For example, Spearman's rank correlation coefficient between final validation accuracy and our best zero-cost proxy on NAS-Bench-201 is 0.82, compared to 0.61 for EcoNAS (a recently proposed reduced-training proxy). Finally, we use these zero-cost proxies to enhance existing NAS search algorithms such as random search, reinforcement learning, evolutionary search and predictor-based search. For all search methodologies and across three different NAS datasets, we are able to significantly improve sample efficiency, and thereby decrease computation, by using our zero-cost proxies. For example on NAS-Bench-101, we achieved the same accuracy 4$\times$ quicker than the best previous result.
翻译:神经结构搜索(NAS)正在迅速成为设计神经网络模型的标准方法。 但是,NAS通常需要大量计算,因为在选择最佳模型之前需要对多个模型进行评估。 为了减少计算力和时间,通常使用代理任务来评价每个模型,而不是全面培训。 在本文中,我们评估常规的减少培训代理,并量化在搜索期间,与最终培训准确度所生成的排名相比,它们保持多种模型之间的排名有多好; 我们根据最近的修剪文献,提议了一系列零成本代理数据,仅仅使用一小批培训数据来计算模型的得分。 为了减少计算量,我们的零成本代理数据需要使用3个数量级的计算,但可以匹配甚至超过常规代理。例如,Spearman的等级相关系数在最终验证准确性与我们在NAS-Bench-201上的最佳零成本代理值之间为0.82,而EcoNAS(最近提议减少培训代理)则为0.61。 最后,我们使用这些零成本代理数据搜索来增强现有的NAS 101 快速搜索效率,我们通过随机搜索和测试来大幅改进现有3个测试。