Neural Architecture Search (NAS) has shown great potentials in finding better neural network designs. Sample-based NAS is the most reliable approach which aims at exploring the search space and evaluating the most promising architectures. However, it is computationally very costly. As a remedy, the one-shot approach has emerged as a popular technique for accelerating NAS using weight-sharing. However, due to the weight-sharing of vastly different networks, the one-shot approach is less reliable than the sample-based approach. In this work, we propose BONAS (Bayesian Optimized Neural Architecture Search), a sample-based NAS framework which is accelerated using weight-sharing to evaluate multiple related architectures simultaneously. Specifically, we apply Graph Convolutional Network predictor as a surrogate model for Bayesian Optimization to select multiple related candidate models in each iteration. We then apply weight-sharing to train multiple candidate models simultaneously. This approach not only accelerates the traditional sample-based approach significantly, but also keeps its reliability. This is because weight-sharing among related architectures are more reliable than those in the one-shot approach. Extensive experiments are conducted to verify the effectiveness of our method over many competing algorithms.
翻译:在寻找更好的神经网络设计方面,以抽样为基础的NAS是旨在探索搜索空间和评估最有希望的建筑的最可靠方法。然而,这是计算成本极高的。作为一种补救办法,一发办法已成为利用权重共享加速NAS的流行技术。然而,由于各大不同网络的权重共享,一发办法比以抽样为基础的方法更不可靠。在这项工作中,我们提议BONAS(Bayesian优化神经建筑搜索),一个以样本为基础的NAS框架,这个框架正在加速使用权重共享来同时评价多个相关结构。具体地说,我们应用“图动网络预测器”作为Bayesian Oppimiz化的替代模型,在每次循环中选择多个相关候选模型。我们随后将权重共享用于同时培训多个候选模型。这个方法不仅大大加快了传统的基于样本的方法,而且还保持了其可靠性。这是因为相关结构之间的权重共享比在一发式算法中的方法更加可靠。我们进行了广泛的实验,以多种方法来进行。