Recently, there has been a growing interest in automating the process of neural architecture design, and the Differentiable Architecture Search (DARTS) method makes the process available within a few GPU days. However, the performance of DARTS is often observed to collapse when the number of search epochs becomes large. Meanwhile, lots of "{\em skip-connect}s" are found in the selected architectures. In this paper, we claim that the cause of the collapse is that there exists overfitting in the optimization of DARTS. Therefore, we propose a simple and effective algorithm, named "DARTS+", to avoid the collapse and improve the original DARTS, by "early stopping" the search procedure when meeting a certain criterion. We also conduct comprehensive experiments on benchmark datasets and different search spaces and show the effectiveness of our DARTS+ algorithm, and DARTS+ achieves $2.32\%$ test error on CIFAR10, $14.87\%$ on CIFAR100, and $23.7\%$ on ImageNet. We further remark that the idea of "early stopping" is implicitly included in some existing DARTS variants by manually setting a small number of search epochs, while we give an {\em explicit} criterion for "early stopping".
翻译:最近,人们对神经结构设计过程自动化的兴趣日益浓厚,而不同的建筑搜索(DARTS)方法使这一过程在数个GPU日内可以使用。然而,当搜索时代数目大时,DARSS的性能经常被观察到崩溃。与此同时,在选定的建筑中发现了许多“它们跳过连接”的功能。在本文中,我们声称,崩溃的原因是DARSS的优化存在过度的功能。因此,我们提议了一个简单有效的算法,名为“DARTS+”,以避免崩溃,改进最初的DARSS,方法是在达到某一标准时“及早停止”搜索程序。我们还在基准数据集和不同的搜索空间上进行了全面实验,显示了我们的DARSS+算法的有效性。DARSS+在CIFAR10、CIFAR100、CIFAR100和图像Net上实现了2.37$的测试错误。我们进一步说,“早期停止”的概念隐含在某个现有的DARS标准中,给我们一个明确的搜索标准。