Bayesian optimization (BO) has become an established framework and popular tool for hyperparameter optimization (HPO) of machine learning (ML) algorithms. While known for its sample-efficiency, vanilla BO can not utilize readily available prior beliefs the practitioner has on the potential location of the optimum. Thus, BO disregards a valuable source of information, reducing its appeal to ML practitioners. To address this issue, we propose $\pi$BO, an acquisition function generalization which incorporates prior beliefs about the location of the optimum in the form of a probability distribution, provided by the user. In contrast to previous approaches, $\pi$BO is conceptually simple and can easily be integrated with existing libraries and many acquisition functions. We provide regret bounds when $\pi$BO is applied to the common Expected Improvement acquisition function and prove convergence at regular rates independently of the prior. Further, our experiments show that $\pi$BO outperforms competing approaches across a wide suite of benchmarks and prior characteristics. We also demonstrate that $\pi$BO improves on the state-of-the-art performance for a popular deep learning task, with a 12.5 $\times$ time-to-accuracy speedup over prominent BO approaches.
翻译:Bayesian优化(BO)已成为机械学习算法超参数优化(HPO)的既定框架和流行工具,尽管以其抽样效率而著称,但Vanilla Bos不能利用执业者对最佳选择的潜在位置的现成先前的信念。因此,BO忽视了宝贵的信息来源,减少了其对ML执业者的吸引力。为了解决这一问题,我们提议采用“Pi$BO”这一购置功能,将用户先前对以概率分布形式表示的最佳地点的信念包含在内。与以往的做法不同,$\pi$BO在概念上是简单的,很容易与现有图书馆和许多购置功能融合。当将$\pi$BO应用于共同的预期改进获取功能时,我们提供了遗憾的界限,并证明以正常的费率与以前不同。此外,我们的实验表明,$pi$BO在广泛的基准和先前特点中,超越了相互竞争的方法。我们还表明,美元BOOO在公众深层次学习任务方面,在州-艺术业绩上有所改进,具有12.5美元至时间的显著速度方法。