Despite the success of recent Neural Architecture Search (NAS) methods on various tasks which have shown to output networks that largely outperform human-designed networks, conventional NAS methods have mostly tackled the optimization of searching for the network architecture for a single task (dataset), which does not generalize well across multiple tasks (datasets). Moreover, since such task-specific methods search for a neural architecture from scratch for every given task, they incur a large computational cost, which is problematic when the time and monetary budget are limited. In this paper, we propose an efficient NAS framework that is trained once on a database consisting of datasets and pretrained networks and can rapidly search for a neural architecture for a novel dataset. The proposed MetaD2A (Meta Dataset-to-Architecture) model can stochastically generate graphs (architectures) from a given set (dataset) via a cross-modal latent space learned with amortized meta-learning. Moreover, we also propose a meta-performance predictor to estimate and select the best architecture without direct training on target datasets. The experimental results demonstrate that our model meta-learned on subsets of ImageNet-1K and architectures from NAS-Bench 201 search space successfully generalizes to multiple unseen datasets including CIFAR-10 and CIFAR-100, with an average search time of 33 GPU seconds. Even under MobileNetV3 search space, MetaD2A is 5.5K times faster than NSGANetV2, a transferable NAS method, with comparable performance. We believe that the MetaD2A proposes a new research direction for rapid NAS as well as ways to utilize the knowledge from rich databases of datasets and architectures accumulated over the past years. Code is available at https://github.com/HayeonLee/MetaD2A.
翻译:尽管最近的神经结构搜索(NAS)方法在各种任务上取得了成功,这些方法向产出网络展示了效率更高的NAS框架,这些方法基本上优于人设计的网络,但常规NAS方法大多解决了为单一任务(数据集)搜索网络架构的最优化,该任务(数据集)没有在多个任务(数据集)中一刀二刀地对每个任务(数据集)进行。此外,由于这些特定任务特定方法从头开始对神经结构进行搜索,因此在时间和货币预算有限的情况下,这些计算成本很高。在本文中,我们提议了一个高效的NAS框架,这个框架曾经在由数据集和预培训网络组成的数据库中受过过培训,并且能够迅速为新数据集(数据集)寻找神经结构架构。拟议的MetaD2A(Meta DS-HA-HAR)模型能够从一个特定的数据集(图表)从一个跨模式的神经结构(数据集)到一个MARSA-DO级的模型。我们还提议用元性预测器来估算和选择最佳的架构,而无需直接在目标数据集中进行直接培训。