Thompson Sampling (TS) from Gaussian Process (GP) models is a powerful tool for the optimization of black-box functions. Although TS enjoys strong theoretical guarantees and convincing empirical performance, it incurs a large computational overhead that scales polynomially with the optimization budget. Recently, scalable TS methods based on sparse GP models have been proposed to increase the scope of TS, enabling its application to problems that are sufficiently multi-modal, noisy or combinatorial to require more than a few hundred evaluations to be solved. However, the approximation error introduced by sparse GPs invalidates all existing regret bounds. In this work, we perform a theoretical and empirical analysis of scalable TS. We provide theoretical guarantees and show that the drastic reduction in computational complexity of scalable TS can be enjoyed without loss in the regret performance over the standard TS. These conceptual claims are validated for practical implementations of scalable TS on synthetic benchmarks and as part of a real-world high-throughput molecular design task.
翻译:Gaussian Process(GP) 模型的Thompson Sampling (TS) 是优化黑盒功能的有力工具。 虽然TS享有强大的理论保障和令人信服的实证性能,但它产生了一个庞大的计算间接成本,以通过优化预算来实现多元规模。 最近,基于稀有的GP模型的可缩放的TS方法被提出来扩大TS的范围,使TS能够应用到足够多模式、吵闹或组合的问题上,需要数百次以上的评估才能解决。然而,由稀有的GPS引入的近距离错误使现有的所有遗憾界限失效。 在这项工作中,我们对可缩放的TS进行理论和经验分析。我们提供了理论保证,并表明,可缩放的TS的计算复杂性可以在不丧失标准TS的遗憾性表现的情况下得到大幅度降低。这些概念性主张得到验证,以实际执行可缩放的合成基准和作为真实世界高载量分子设计任务的一部分。