Differentiable architecture search (DARTS) marks a milestone in Neural Architecture Search (NAS), boasting simplicity and small search costs. However, DARTS still suffers from frequent performance collapse, which happens when some operations, such as skip connections, zeroes and poolings, dominate the architecture. In this paper, we are the first to point out that the phenomenon is attributed to bi-level optimization. We propose Single-DARTS which merely uses single-level optimization, updating network weights and architecture parameters simultaneously with the same data batch. Even single-level optimization has been previously attempted, no literature provides a systematic explanation on this essential point. Replacing the bi-level optimization, Single-DARTS obviously alleviates performance collapse as well as enhances the stability of architecture search. Experiment results show that Single-DARTS achieves state-of-the-art performance on mainstream search spaces. For instance, on NAS-Benchmark-201, the searched architectures are nearly optimal ones. We also validate that the single-level optimization framework is much more stable than the bi-level one. We hope that this simple yet effective method will give some insights on differential architecture search. The code is available at https://github.com/PencilAndBike/Single-DARTS.git.
翻译:可区别的建筑搜索(DARTS)标志着神经结构搜索(NAS)的一个里程碑,它自夸简洁和少量搜索成本。然而,DARTS仍然经常出现性能崩溃,这种情况发生在一些操作,例如跳过连接、零和集合,主宰了建筑。在本文中,我们首先指出,这种现象是双级优化造成的。我们建议仅使用单级优化、更新网络重量和架构参数与同一数据批次同时使用的单级DARTS。即使以前曾尝试过单级优化,也没有文献对此基本点提供系统的解释。替换双级优化,单级数据系统显然减轻了性能崩溃,并加强了建筑搜索的稳定性。实验结果显示,单级数据搜索系统在主流搜索空间达到最新水平的性能。例如,在NAS-Benchmark-201上,搜索的架构几乎是最佳的。我们还验证了单级优化框架比双级框架更稳定得多。我们希望这一简单有效的方法能减轻性绩效崩溃,并且能够增强建筑搜索的稳定性。在 MASGI/AGIKS。