Automated machine learning (AutoML) usually involves several crucial components, such as Data Augmentation (DA) policy, Hyper-Parameter Optimization (HPO), and Neural Architecture Search (NAS). Although many strategies have been developed for automating these components in separation, joint optimization of these components remains challenging due to the largely increased search dimension and the variant input types of each component. In parallel to this, the common practice of searching for the optimal architecture first and then retraining it before deployment in NAS often suffers from low performance correlation between the searching and retraining stages. An end-to-end solution that integrates the AutoML components and returns a ready-to-use model at the end of the search is desirable. In view of these, we propose DHA, which achieves joint optimization of Data augmentation policy, Hyper-parameter and Architecture. Specifically, end-to-end NAS is achieved in a differentiable manner by optimizing a compressed lower-dimensional feature space, while DA policy and HPO are regarded as dynamic schedulers, which adapt themselves to the update of network parameters and network architecture at the same time. Experiments show that DHA achieves state-of-the-art (SOTA) results on various datasets and search spaces. To the best of our knowledge, we are the first to efficiently and jointly optimize DA policy, NAS, and HPO in an end-to-end manner without retraining.
翻译:自动机器学习(Automal)通常涉及几个关键组成部分,如数据增强(DA)政策、超光谱优化(HPO)和神经结构搜索(NAS)等数据增强(Automal)政策。虽然已经制定了许多战略,将这些组成部分分离成自动化,但是由于搜索层面和每个组成部分的变式输入类型大大增加,这些组成部分的联合优化仍然具有挑战性。与此相平行,首先寻找最佳架构然后在NAS部署之前再再培训这一共同做法往往因搜索和再培训阶段之间的性能相关性低而受到影响。在搜索结束时,采用将自动MLE组件整合并返回一个随时使用的模型的端对端解决方案是可取的。有鉴于此,我们建议DHA, 实现数据增强政策、超光度和结构的联合优化。具体地说,端对端到端NAS的实现不同方式,优化的地地地表空间,而DA政策和HO被视为动态的调度器,在搜索结束时根据网络参数和网络结构的更新情况进行调整。实验显示,我们的最佳搜索方式是在不同时进行最佳搜索。