Differentiable ARchiTecture Search (DARTS) is one of the most trending Neural Architecture Search (NAS) methods, drastically reducing search cost by resorting to Stochastic Gradient Descent (SGD) and weight-sharing. However, it also greatly reduces the search space, thus excluding potential promising architectures from being discovered. In this paper, we propose D-DARTS, a novel solution that addresses this problem by nesting several neural networks at cell-level instead of using weight-sharing to produce more diversified and specialized architectures. Moreover, we introduce a novel algorithm which can derive deeper architectures from a few trained cells, increasing performance and saving computation time. Our solution is able to provide state-of-the-art results on CIFAR-10, CIFAR-100 and ImageNet while using significantly less parameters than previous baselines, resulting in more hardware-efficient neural networks.
翻译:可区别的ARCHTITETS(DARTS)是最具趋势的神经结构搜索(NAS)方法之一,通过采用Stochatic Gradient Emple(SGD)和权重共享,大幅降低了搜索成本。然而,这也大大减少了搜索空间,从而将潜在的有希望的建筑排除在被发现之外。在本文中,我们提出D-DARTS,这是一个新颖的解决办法,通过在细胞一级嵌入多个神经网络,而不是利用权重共享来生产更加多样化和专业化的建筑。此外,我们引入了一种新颖的算法,它可以从少数受过训练的细胞中获取更深层的建筑,增加性能和节省计算时间。我们的解决方案能够提供CIFAR-10、CIFAR-100和图像网络的最新结果,同时使用比以前的基线要少得多的参数,从而导致硬件效率更高的神经网络。