In addition to the best model architecture and hyperparameters, a full AutoML solution requires selecting appropriate hardware automatically. This can be framed as a multi-objective optimization problem: there is not a single best hardware configuration but a set of optimal ones achieving different trade-offs between cost and runtime. In practice, some choices may be overly costly or take days to train. To lift this burden, we adopt a multi-objective approach that selects and adapts the hardware configuration automatically alongside neural architectures and their hyperparameters. Our method builds on Hyperband and extends it in two ways. First, we replace the stopping rule used in Hyperband by a non-dominated sorting rule to preemptively stop unpromising configurations. Second, we leverage hyperparameter evaluations from related tasks via transfer learning by building a probabilistic estimate of the Pareto front that finds promising configurations more efficiently than random search. We show in extensive NAS and HPO experiments that both ingredients bring significant speed-ups and cost savings, with little to no impact on accuracy. In three benchmarks where hardware is selected in addition to hyperparameters, we obtain runtime and cost reductions of at least 5.8x and 8.8x, respectively. Furthermore, when applying our multi-objective method to the tuning of hyperparameters only, we obtain a 10\% improvement in runtime while maintaining the same accuracy on two popular NAS benchmarks.
翻译:除了最佳模型架构和超参数外,一个完整的自动解剖解决方案要求自动选择适当的硬件。它可以被设计成一个多目标优化问题:没有单一的最佳硬件配置,而是一套最佳的配置,在成本和运行时间之间实现不同的权衡。实际上,一些选择可能费用过高,或者需要花费几天时间来培训。为了减轻这一负担,我们采取了一个多目标方法,在神经结构及其超光谱结构及其超光谱中自动选择和调整硬件配置。我们的方法建立在超光速波段上,并且以两种方式扩展。首先,我们用非主导性排序规则取代超音频带中使用的停用规则,先先行停止不促进配置。第二,我们利用超光谱评价从相关任务中进行转移,方法是对Pareto前方进行预测性估计,发现有希望的配置比随机搜索更有效。我们通过广泛的NAS和HPO实验显示,这两个要素都带来显著的速度和成本节约,对准确性影响不大。在三个基准中,除超光度外,我们选择了超频段排序规则,我们只能使用运行时间和最精确度标准,在10x级标准标准上分别调整。