Investigating uncertainties in computer simulations can be prohibitive in terms of computational costs, since the simulator needs to be run over a large number of input values. Building an emulator, i.e. a statistical surrogate model of the simulator constructed using a design of experiments made of a comparatively small number of evaluations of the forward solver, greatly alleviates the computational burden to carry out such investigations. Nevertheless, this can still be above the computational budget for many studies. Two major approaches have been used to reduce the budget needed to build the emulator: efficient design of experiments, such as sequential designs, and combining training data of different degrees of sophistication in a so-called multi-fidelity method, or multilevel in case these fidelities are ordered typically for increasing resolutions. We present here a novel method that combines both approaches, the multilevel adaptive sequential design of computer experiments (MLASCE) in the framework of Gaussian process (GP) emulators. We make use of reproducing kernel Hilbert spaces as a tool for our GP approximations of the differences between two consecutive levels. This dual strategy allows us to allocate efficiently limited computational resources over simulations of different levels of fidelity and build the GP emulator. The allocation of computational resources is shown to be the solution of a simple optimization problem in a special case where we theoretically prove the validity of our approach. Our proposed method is compared with other existing models of multi-fidelity Gaussian process emulation. Gains in orders of magnitudes in accuracy or computing budgets are demonstrated in some of numerical examples for some settings.
翻译:计算机模拟的不确定性在计算成本方面可能令人望而却步,因为模拟器需要用大量输入值来运行。 建立一个模拟器, 即一个模拟器的统计替代模型, 模拟器是使用对远端求解器进行数量相对较少的评价的实验设计的模拟器, 大大减轻了进行这种调查的计算负担。 然而, 这仍然可能超出许多研究的计算预算。 已经使用了两大方法来减少建立模拟器所需的预算: 高效设计实验, 如顺序设计, 并结合不同程度的精密培训数据, 即所谓的多纤维化方法, 或多层次的模拟器, 通常为了增加分辨率而设置这些真实性。 我们在此提出了一个新颖的方法, 将两种方法结合起来, 多层次的适应性序列设计, 进行这样的计算机实验, 用于高斯进程( GPG) 模擬器框架。 我们利用再生内流希尔伯特空间作为工具, 用于我们GPG的精确度, 如顺序设计, 将不同程度的培训数据合并在一起, 或者多层次的计算, 在两个连续的计算中, 这个双重的计算方法中, 显示我们的资源的模拟中, 能够有效地分配。