Neural Architecture Search (NAS) is an automatic technique that can search for well-performed architectures for a specific task. Although NAS surpasses human-designed architecture in many fields, the high computational cost of architecture evaluation it requires hinders its development. A feasible solution is to directly evaluate some metrics in the initial stage of the architecture without any training. NAS without training (WOT) score is such a metric, which estimates the final trained accuracy of the architecture through the ability to distinguish different inputs in the activation layer. However, WOT score is not an atomic metric, meaning that it does not represent a fundamental indicator of the architecture. The contributions of this paper are in three folds. First, we decouple WOT into two atomic metrics which represent the distinguishing ability of the network and the number of activation units, and explore better combination rules named (Distinguishing Activation Score) DAS. We prove the correctness of decoupling theoretically and confirmed the effectiveness of the rules experimentally. Second, in order to improve the prediction accuracy of DAS to meet practical search requirements, we propose a fast training strategy. When DAS is used in combination with the fast training strategy, it yields more improvements. Third, we propose a dataset called Darts-training-bench (DTB), which fills the gap that no training states of architecture in existing datasets. Our proposed method has 1.04$\times$ - 1.56$\times$ improvements on NAS-Bench-101, Network Design Spaces, and the proposed DTB.
翻译:神经结构搜索(NAS)是一种自动技术,可以为某项具体任务寻找完善的建筑。尽管NAS在许多领域超越了人设计的建筑,但建筑评估的计算成本很高,这阻碍了建筑评估的发展。一个可行的解决办法是直接评价建筑初始阶段的一些测量标准,而没有任何培训。没有培训的NAS是这样一个衡量标准,它通过区分激活层不同投入的能力来估计该建筑经过培训的最终准确性。然而,WOT的得分不是原子指标,这意味着它不代表建筑的基本指标。本文的贡献是三个折叠的。首先,我们将WOT分解成两个原子计量标准,代表网络和启动单位的数量的区别性能力,并探索更好的组合规则(DOTS动作评分) DAS。我们证明从理论上解析的正确性,并证实了规则的有效性。第二,为了提高DAS的预测准确性,以达到实际搜索要求,我们提议了一个快速培训战略。当DAS与我们快速使用数据组合时,DAS将数据组合起来,我们建议了一个快速培训方法。