Hyperparameter optimization (HPO) and neural architecture search (NAS) are methods of choice to obtain the best-in-class machine learning models, but in practice they can be costly to run. When models are trained on large datasets, tuning them with HPO or NAS rapidly becomes prohibitively expensive for practitioners, even when efficient multi-fidelity methods are employed. We propose an approach to tackle the challenge of tuning machine learning models trained on large datasets with limited computational resources. Our approach, named PASHA, extends ASHA and is able to dynamically allocate maximum resources for the tuning procedure depending on the need. The experimental comparison shows that PASHA identifies well-performing hyperparameter configurations and architectures while consuming significantly fewer computational resources than ASHA.
翻译:超光计优化和神经结构搜索(NAS)是获得最优级机床学习模型的首选方法,但实际上它们运行成本可能很高。当模型在大型数据集上接受培训时,使用HPO或NAS对模型进行调试,对实践者来说成本非常高,即使采用了高效的多纤维方法。我们建议了一种方法来应对在有限的计算资源范围内对大型数据集培训的机器学习模型进行调试的挑战。我们称为PASHA的方法扩展了ASHA,能够根据需要动态地为调试程序分配最大资源。实验性比较表明,PASHA在消耗比ASHA少得多的计算资源的同时,确定了良好的超参数配置和结构。</s>