Neural architecture search (NAS) has attracted a lot of attention and has been illustrated to bring tangible benefits in a large number of applications in the past few years. Architecture topology and architecture size have been regarded as two of the most important aspects for the performance of deep learning models and the community has spawned lots of searching algorithms for both aspects of the neural architectures. However, the performance gain from these searching algorithms is achieved under different search spaces and training setups. This makes the overall performance of the algorithms to some extent incomparable and the improvement from a sub-module of the searching model unclear. In this paper, we propose NATS-Bench, a unified benchmark on searching for both topology and size, for (almost) any up-to-date NAS algorithm. NATS-Bench includes the search space of 15,625 neural cell candidates for architecture topology and 32,768 for architecture size on three datasets. We analyze the validity of our benchmark in terms of various criteria and performance comparison of all candidates in the search space. We also show the versatility of NATS-Bench by benchmarking 13 recent state-of-the-art NAS algorithms on it. All logs and diagnostic information trained using the same setup for each candidate are provided. This facilitates a much larger community of researchers to focus on developing better NAS algorithms in a more comparable and computationally cost friendly environment. All codes are publicly available at: https://xuanyidong.com/assets/projects/NATS-Bench.
翻译:过去几年来,建筑表层和建筑规模被视为深层学习模型表现的两个最重要的方面,社区为神经结构的这两个方面产生了大量的搜索算法。然而,这些搜索算法的绩效是在不同搜索空间和培训设置下取得的。这使得算法的整体性能在某种程度上无法比较,而且搜索模型的一个子模块的改进情况也不明确。在本论文中,我们提议NATS-Bench,一个搜索表层和规模的统一基准,用于(几乎)任何最新的NAS算法。NATS-Bench包含大量搜索神经结构结构两个方面的搜索算法。但是,从这些搜索算法中获得的绩效是在不同搜索空间和训练设置的搜索空间和训练结构大小的15,625个神经细胞候选人的搜索空间,32,768个。我们从各种标准的角度分析了我们基准的可靠性以及搜索空间所有候选人的绩效比较。我们还提出了NATS-Bench在13个层次和(最新的)社区查找规模上的统一基准。通过对13个州/候选人的高级分析算法进行更精确的测试,在13个州/候选人的逻辑上提供了更精确的逻辑分析。