In recent years, neural architecture search (NAS) methods have been proposed for the automatic generation of task-oriented network architecture in image classification. However, the architectures obtained by existing NAS approaches are optimized only for classification performance and do not adapt to devices with limited computational resources. To address this challenge, we propose a neural network architecture search algorithm aiming to simultaneously improve network performance (e.g., classification accuracy) and reduce network complexity. The proposed framework automatically builds the network architecture at two stages: block-level search and network-level search. At the stage of block-level search, a relaxation method based on the gradient is proposed, using an enhanced gradient to design high-performance and low-complexity blocks. At the stage of network-level search, we apply an evolutionary multi-objective algorithm to complete the automatic design from blocks to the target network. The experiment results demonstrate that our method outperforms all evaluated hand-crafted networks in image classification, with an error rate of on CIFAR10 and an error rate of on CIFAR100, both at network parameter size less than one megabit. Moreover, compared with other neural architecture search methods, our method offers a tremendous reduction in designed network architecture parameters.
翻译:近年来,为在图像分类中自动生成以任务为导向的网络架构,提出了神经结构搜索(NAS)方法,目的是在图像分类中自动生成以任务为导向的网络架构;然而,现有NAS方法获得的架构仅优化用于分类性能,不适应计算资源有限的装置;为了应对这一挑战,我们提议了一个神经网络结构搜索算法,目的是同时改善网络性能(例如分类精确度)和降低网络复杂性;拟议框架在两个阶段自动建立网络架构:区块级搜索和网络级搜索;在区块级搜索阶段,提出了基于梯度的放松方法,使用强化梯度设计高性能和低兼容度区块;在网络搜索阶段,我们采用进化的多客观算法完成从区块到目标网络的自动设计;实验结果表明,我们的方法超越了图像分类中所有经评价的手工艺网络,在CIFAR10上的误差率和CIFAR100的误差率,两者在网络参数大小小于一个兆位上都采用了一个梯度。此外,与其他神经结构搜索结构的参数相比,我们的方法提供了巨大的缩小。