To achieve peak predictive performance, hyperparameter optimization (HPO) is a crucial component of machine learning and its applications. Over the last years, the number of efficient algorithms and tools for HPO grew substantially. At the same time, the community is still lacking realistic, diverse, computationally cheap, and standardized benchmarks. This is especially the case for multi-fidelity HPO methods. To close this gap, we propose HPOBench, which includes 7 existing and 5 new benchmark families, with a total of more than 100 multi-fidelity benchmark problems. HPOBench allows to run this extendable set of multi-fidelity HPO benchmarks in a reproducible way by isolating and packaging the individual benchmarks in containers. It also provides surrogate and tabular benchmarks for computationally affordable yet statistically sound evaluations. To demonstrate HPOBench's broad compatibility with various optimization tools, as well as its usefulness, we conduct an exemplary large-scale study evaluating 13 optimizers from 6 optimization tools. We provide HPOBench here: https://github.com/automl/HPOBench.
翻译:为了达到高峰预测性能,超光谱优化(HPO)是机器学习及其应用的一个关键组成部分。过去几年来,HPO的有效算法和工具数量大幅增长。与此同时,社区仍然缺乏现实、多样化、计算上便宜和标准化的基准。特别是多信仰的HPO方法。为了缩小这一差距,我们提议HPOBench,它包括7个现有和5个新的基准家庭,共有100多个多信仰基准问题。HPOBench允许通过隔离和包装集装箱中的单个基准,以可复制的方式运行这套可扩展的多信仰HPO基准。它还为计算上可负担得起但具有统计性的评价提供了代理和表格基准。为了证明HPOBench与各种优化工具的广泛兼容性及其效用,我们进行了一项模拟性的大规模研究,从6个优化工具中评估13个优化器。我们在这里提供HPOBench:https://github.com/autooml/HPOBebench。