Recent neural architecture search (NAS) works proposed training-free metrics to rank networks which largely reduced the search cost in NAS. In this paper, we revisit these training-free metrics and find that: (1) the number of parameters (\#Param), which is the most straightforward training-free metric, is overlooked in previous works but is surprisingly effective, (2) recent training-free metrics largely rely on the \#Param information to rank networks. Our experiments show that the performance of recent training-free metrics drops dramatically when the \#Param information is not available. Motivated by these observations, we argue that metrics less correlated with the \#Param are desired to provide additional information for NAS. We propose a light-weight training-based metric which has a weak correlation with the \#Param while achieving better performance than training-free metrics at a lower search cost. Specifically, on DARTS search space, our method completes searching directly on ImageNet in only 2.6 GPU hours and achieves a top-1/top-5 error rate of 24.1\%/7.1\%, which is competitive among state-of-the-art NAS methods. Codes are available at \url{https://github.com/taoyang1122/Revisit_TrainingFree_NAS}
翻译:最近,神经结构搜索(NAS)为网络排级提出了无培训指标,这大大降低了NAS的搜索成本。在本文中,我们重新审视了这些无培训指标,发现:(1) 参数数量( ⁇ Param),这是最直接的无培训指标,在以前的工作中被忽视,但令人惊讶的是效果;(2) 最近无培训指标,主要依靠 ⁇ Param信息对网络进行排级;我们的实验表明,在没有提供“Param”信息的情况下,最近无培训指标的运行量急剧下降。根据这些观察,我们争辩说,与“Param”不相干的指标希望为NAS提供更多的信息。我们建议采用与“Param”的轻量级培训标准,该标准与“Param”的相关性较弱,同时在较低的搜索成本下比无培训指标的绩效要好。具体地说,在DARTS搜索空间,我们的方法在仅2.6个GPU小时的图像网络上直接完成搜索,并达到24.1/top-5的错误率,即24.1 ⁇ /Rev.1 ⁇,该标准在州-Art-ausfrema_ar_ar_NASqistrium_Amb可使用的方法。