Bayesian optimization (BO) has been widely used in machine learning and simulation optimization. With the increase in computational resources and storage capacities in these fields, high-dimensional and large-scale problems are becoming increasingly common. In this study, we propose a model aggregation method in the Bayesian optimization (MamBO) algorithm for efficiently solving high-dimensional large-scale optimization problems. MamBO uses a combination of subsampling and subspace embeddings to collectively address high dimensionality and large-scale issues; in addition, a model aggregation method is employed to address the surrogate model uncertainty issue that arises when embedding is applied. This surrogate model uncertainty issue is largely ignored in the embedding literature and practice, and it is exacerbated when the problem is high-dimensional and data are limited. Our proposed model aggregation method reduces these lower-dimensional surrogate model risks and improves the robustness of the BO algorithm. We derive an asymptotic bound for the proposed aggregated surrogate model and prove the convergence of MamBO. Benchmark numerical experiments indicate that our algorithm achieves superior or comparable performance to other commonly used high-dimensional BO algorithms. Moreover, we apply MamBO to a cascade classifier of a machine learning algorithm for face detection, and the results reveal that MamBO finds settings that achieve higher classification accuracy than the benchmark settings and is computationally faster than other high-dimensional BO algorithms.
翻译:Bayesian优化(BO)已被广泛用于机器学习和模拟优化(BO) 。随着这些领域计算资源和存储能力的增加,高维和大规模问题日益普遍。在本研究中,我们提议在Bayesian优化(MambO)算法中采用模型汇总方法,以高效解决高维大型优化问题。MambO采用子取样和子空间嵌入组合组合,以集体解决高维性和大规模问题;此外,还采用模型汇总方法来解决在嵌入时产生的代金模型不确定性问题。这种代金模型不确定性问题在嵌入文献和实践中基本上被忽视,而当问题为高维和数据有限时,这一问题就更加严重。我们提议的模型汇总方法减少了这些低维代金模型风险,提高了BO算的稳健性。我们为拟议的总超维基模型模型模型模型模型和证明MamBO的面面的趋同性。基准实验表明,我们的替代模型的性能优或可比性业绩与其他常用的高维基BO的代算法相比,我们用了一个高级BA级的升级的升级和升级的机级标准。此外,BO的升级的升级的系统的升级的升级的升级的升级的升级和升级的升级的升级的升级是机器的升级的升级。