Modern Neural Architecture Search methods have repeatedly broken state-of-the-art results for several disciplines. The super-network, a central component of many such methods, enables quick estimates of accuracy or loss statistics for any architecture in the search space. They incorporate the network weights of all candidate architectures and can thus approximate specific ones by applying the respective operations. However, this design ignores potential dependencies between consecutive operations. We extend super-networks with conditional weights that depend on combinations of choices and analyze their effect. Experiments in NAS-Bench 201 and NAS-Bench-Macro-based search spaces show improvements in the architecture selection and that the resource overhead is nearly negligible for sequential network designs.
翻译:现代神经架构搜索方法一再打破几个学科的最新成果。超级网络是许多此类方法的核心组成部分,它能够快速估计搜索空间中任何建筑的准确性或损失统计数据,它们包含所有候选建筑的网络权重,因此可以通过应用各自的操作来接近特定结构。然而,这一设计忽视了连续运行之间潜在的依赖性。我们扩展了有条件的超网络,这些网络取决于选择组合并分析其效果。NAS-Bench 201和NAS-Bench-Macro搜索空间的实验显示,建筑选择有所改善,而且连续网络设计的资源间接费用几乎微不足道。