Neural Architecture Search (NAS) can automatically design architectures for deep neural networks (DNNs) and has become one of the hottest research topics in the current machine learning community. However, NAS is often computationally expensive because a large number of DNNs require to be trained for obtaining performance during the search process. Performance predictors can greatly alleviate the prohibitive cost of NAS by directly predicting the performance of DNNs. However, building satisfactory performance predictors highly depends on enough trained DNN architectures, which are difficult to obtain in most scenarios. To solve this critical issue, we propose an effective DNN architecture augmentation method named GIAug in this paper. Specifically, we first propose a mechanism based on graph isomorphism, which has the merit of efficiently generating a factorial of $\boldsymbol n$ (i.e., $\boldsymbol n!$) diverse annotated architectures upon a single architecture having $\boldsymbol n$ nodes. In addition, we also design a generic method to encode the architectures into the form suitable to most prediction models. As a result, GIAug can be flexibly utilized by various existing performance predictors-based NAS algorithms. We perform extensive experiments on CIFAR-10 and ImageNet benchmark datasets on small-, medium- and large-scale search space. The experiments show that GIAug can significantly enhance the performance of most state-of-the-art peer predictors. In addition, GIAug can save three magnitude order of computation cost at most on ImageNet yet with similar performance when compared with state-of-the-art NAS algorithms.
翻译:神经结构搜索(NAS)可以自动设计深神经网络的架构,并已成为当前机器学习界中最热的研究课题之一。然而,NAS往往计算成本昂贵,因为大量DNS需要经过培训才能在搜索过程中获得性能。 性能预测可以通过直接预测DNS的性能,大大降低NAS令人望而生畏的成本。 然而,建立令人满意的性能预测器高度依赖于经过充分培训的DNNN结构,在多数情况下很难获得。为了解决这一关键问题,我们提出了在本文中名为 GIAug 的有效DNN 结构增强方法。具体地说,我们首先提议基于图形无形态的机械化机制,其优点是高效生成一个要素值为$\boldsymsbol n. 。 性能预测器通过直接预测DNNNN( $\boldsymbol n.) 的单一结构的注解图。此外,我们还可以设计一种通用的方法,将结构编码成适合最小型预测模型的表格。我们首先提出基于图形的图形的图象学形态的图象学的图象化的模型,作为结果,全球网络的模型的模型的模型的模型的模拟的模型的模拟的模拟的模拟的模拟的模型可以大量性性能评估。我们利用了各种的模型的模型的模型的模拟的模拟的模拟的模拟的模拟的模拟的模拟的性能能能。