We formalize and analyze a fundamental component of differentiable neural architecture search (NAS): local "operation scoring" at each operation choice. We view existing operation scoring functions as inexact proxies for accuracy, and we find that they perform poorly when analyzed empirically on NAS benchmarks. From this perspective, we introduce a novel \textit{perturbation-based zero-cost operation scoring} (Zero-Cost-PT) approach, which utilizes zero-cost proxies that were recently studied in multi-trial NAS but degrade significantly on larger search spaces, typical for differentiable NAS. We conduct a thorough empirical evaluation on a number of NAS benchmarks and large search spaces, from NAS-Bench-201, NAS-Bench-1Shot1, NAS-Bench-Macro, to DARTS-like and MobileNet-like spaces, showing significant improvements in both search time and accuracy. On the ImageNet classification task on the DARTS search space, our approach improved accuracy compared to the best current training-free methods (TE-NAS) while being over 10$\times$ faster (total searching time 25 minutes on a single GPU), and observed significantly better transferability on architectures searched on the CIFAR-10 dataset with an accuracy increase of 1.8 pp. Our code is available at: https://github.com/zerocostptnas/zerocost_operation_score.
翻译:我们正式确定和分析不同神经结构搜索的基本组成部分:每个作业选择的地方“操作评分”,我们视现有操作评分功能为不精确的替代物,我们发现在对NAS基准进行实证分析时,这些评分功能的表现不尽如人意。从这个角度,我们引入了新型的Textit{turbation-基础零成本操作评分}(零成本-PT)方法,该方法利用了最近在多审NAS中研究的零成本代理物,但在更大的搜索空间中却显著下降,这是不同的NAS典型的典型。我们对现有NAS基准和大型搜索空间进行彻底的经验性评估,从NAS-Bench-201,NAS-Bench-1Shot1,NAS-Bench-1,NAS-Bench-Macro,到DARSS类似和移动网络的空域,在搜索时间和准确性两方面都有显著改进。关于DARSS搜索空间的图像网络分类任务,我们的方法比目前最佳的无培训方法(TE-NAS)的精确度有所提高,同时进行超过10美元/时间的搜索。