Recent advances in Neural Architecture Search (NAS) such as one-shot NAS offer the ability to extract specialized hardware-aware sub-network configurations from a task-specific super-network. While considerable effort has been employed towards improving the first stage, namely, the training of the super-network, the search for derivative high-performing sub-networks is still under-explored. Popular methods decouple the super-network training from the sub-network search and use performance predictors to reduce the computational burden of searching on different hardware platforms. We propose a flexible search framework that automatically and efficiently finds optimal sub-networks that are optimized for different performance metrics and hardware configurations. Specifically, we show how evolutionary algorithms can be paired with lightly trained objective predictors in an iterative cycle to accelerate architecture search in a multi-objective setting for various modalities including machine translation and image classification.
翻译:最近神经结构搜索(NAS)的进展,如一投射的NAS,能够从一个特定任务的超级网络中提取专门的硬件觉察子网络配置。虽然在改进第一阶段,即超级网络培训方面已经付出了相当大的努力,但是对衍生品高性能子网络的搜索仍然未得到充分探讨。大众方法将超级网络培训与子网络搜索和使用性能预测器相分离,以减少在不同硬件平台上搜索的计算负担。我们提出了一个灵活搜索框架,自动有效地找到最佳的子网络,为不同的性能计量和硬件配置提供优化。具体地说,我们展示了如何在迭代周期中将进化算法与经过轻度培训的客观预测器相匹配,以加速在多目标环境中对各种模式的建筑搜索,包括机器翻译和图像分类。