Hyperparameter optimization aims at finding more rapidly and efficiently the best hyperparameters (HPs) of learning models such as neural networks. In this work, we present a new approach called GPBT (Genealogical Population-Based Training), which shares many points with Population-Based Training: our approach outputs a schedule of HPs and updates both weights and HPs in a single run, but brings several novel contributions: the choice of new HPs is made by a modular search algorithm, the search algorithm can search HPs independently for models with different weights and can exploit separately the maximum amount of meaningful information (genealogically-related) from previous HPs evaluations instead of exploiting together all previous HPs evaluations, a variation of early stopping allows a 2-3 fold acceleration at small performance cost. GPBT significantly outperforms all other approaches of HP Optimization, on all supervised learning experiments tested in terms of speed and performances. HPs tuning will become less computationally expensive using our approach, not only in the deep learning field, but potentially for all processes based on iterative optimization.
翻译:超强参数优化旨在更快、更高效地找到神经网络等学习模型的最佳超参数(HPs)。在这项工作中,我们提出了一个名为GPBT(基于人口的培训)的新方法,它与基于人口的培训分享了许多要点:我们的方法产出了一个HP时间表,并一次性更新了加权和惠普,但带来了一些新的贡献:选择新的HP是模块搜索算法,搜索算法可以独立地搜索具有不同重量的模型,并且可以从以往的HPs评价中分离出最大数量的有意义的信息(与基因有关),而不是利用以往的所有HPs评价,早期停止的变异使得可以以小的性能成本加速2-3倍。 GPBT大大超越了所有其他按速度和性能测试的HPOppimiz化方法,在所有受监督的学习实验中,不仅在深层学习领域,而且有可能对所有基于迭层优化的程序进行计算成本较低。