Adequate labeled data and expensive compute resources are the prerequisites for the success of neural architecture search(NAS). It is challenging to apply NAS in meta-learning scenarios with limited compute resources and data. In this paper, an across-task neural architecture search (AT-NAS) is proposed to address the problem through combining gradient-based meta-learning with EA-based NAS to learn over the distribution of tasks. The supernet is learned over an entire set of tasks by meta-learning its weights. Architecture encodes of subnets sampled from the supernet are iteratively adapted by evolutionary algorithms while simultaneously searching for a task-sensitive meta-network. Searched meta-network can be adapted to a novel task via a few learning steps and only costs a little search time. Empirical results show that AT-NAS surpasses the related approaches on few-shot classification accuracy. The performance of AT-NAS on classification benchmarks is comparable to that of models searched from scratch, by adapting the architecture in less than an hour from a 5-GPU-day pretrained meta-network.
翻译:充分标签数据和昂贵的计算资源是神经结构搜索成功的先决条件。 将NAS应用于计算资源和数据有限的元学习情景具有挑战性。 在本文中,建议通过跨任务神经结构搜索(AT-NAS),将基于梯度的元学习与基于EA的NAS结合起来,学习如何分配任务。超级网通过元学习其重量,在整个任务中学习。从超级网取样的子网的建筑编码通过进化算法进行迭接,同时寻找对任务敏感的元网络。 搜索的元网络可以通过几个学习步骤适应一项新任务,只花费一点时间。 实证结果表明,AT-NAS在分类基准方面超越了对几发分类准确性的相关方法。 AT-NAS在分类基准方面的表现与从零到零的模型相似,在不到一小时的时间里对结构进行调整,从5-GPU的预先训练的元网络中抽取。