This work targets designing a principled and unified training-free framework for Neural Architecture Search (NAS), with high performance, low cost, and in-depth interpretation. NAS has been explosively studied to automate the discovery of top-performer neural networks, but suffers from heavy resource consumption and often incurs search bias due to truncated training or approximations. Recent NAS works start to explore indicators that can predict a network's performance without training. However, they either leveraged limited properties of deep networks, or the benefits of their training-free indicators are not applied to more extensive search methods. By rigorous correlation analysis, we present a unified framework to understand and accelerate NAS, by disentangling "TEG" characteristics of searched networks - Trainability, Expressivity, Generalization - all assessed in a training-free manner. The TEG indicators could be scaled up and integrated with various NAS search methods, including both supernet and single-path approaches. Extensive studies validate the effective and efficient guidance from our TEG-NAS framework, leading to both improved search accuracy and over 2.3x reduction in search time cost. Moreover, we visualize search trajectories on three landscapes of "TEG" characteristics, observing that while a good local minimum is easier to find on NAS-Bench-201 given its simple topology, balancing "TEG" characteristics is much harder on the DARTS search space due to its complex landscape geometry. Our code is available at https://github.com/VITA-Group/TEGNAS.
翻译:这项工作目标是为神经结构搜索设计一个原则性和统一的无培训框架,其性能高、成本低、解释深度高。对神经结构搜索进行了爆炸性研究,目的是将发现最前沿神经网络自动化,但资源消耗量大,往往由于培训或近似不足而产生搜索偏差。近期的NAS工作开始探索能够预测网络业绩的指标,但是,它们要么利用深网络的有限特性,要么将其无培训指标的效益应用于更广泛的搜索方法。通过严格的相关分析,我们提出了一个统一框架来理解和加速NAS,方法是使搜索网络的“TEG”特征——可培训性、快性、通用性――不培训方式评估。TEG指标可以扩大并结合各种NAS搜索方法,包括不经过培训的网络超级网络和单一路径方法。广泛研究验证了我们TEG-NAS框架的有效和高效指导,导致搜索准确性提高搜索准确性,在搜索成本方面减少2.3x以上。此外,我们在搜索的地理结构图上进行简单搜索,同时在搜索时,SEG-EG在搜索中发现最易的地平面图上进行搜索。