In this work, we study the approximation of expected values of functional quantities on the solution of a stochastic differential equation (SDE), where we replace the Monte Carlo estimation with the evaluation of a deep neural network. Once the neural network training is done, the evaluation of the resulting approximating function is computationally highly efficient so that using deep neural networks to replace costly Monte Carlo integration is appealing, e.g., for near real-time computations in quantitative finance. However, the drawback of these nowadays widespread ideas lies in the fact that training a suitable neural network is likely to be prohibitive in terms of computational cost. We address this drawback here by introducing a multilevel approach to the training of deep neural networks. More precisely, we combine the deep learning algorithm introduced by Beck et al. with the idea of multilevel Monte Carlo path simulation of Giles. The idea is to train several neural networks, each having a certain approximation quality and computational complexity, with training data computed from so-called level estimators, introduced by Giles. We show that under certain assumptions, the variance in the training process can be reduced by shifting most of the computational workload to training neural nets at coarse levels where producing the training data sets is comparably cheap, whereas training the neural nets corresponding to the fine levels requires only a limited number of training data sets. We formulate a complexity theorem showing that the multilevel idea can indeed reduce computational complexity.
翻译:在这项工作中,我们研究的是,在解决随机差异方程式(SDE)的解决方案中,功能量的预期值的近似值,我们用对深神经网络的评估来取代蒙特卡洛的估算。一旦神经网络培训完成后,对由此产生的近似功能的评价就具有很高的计算效率,因此,利用深神经网络来取代昂贵的蒙特卡洛整合的构想就具有吸引力,例如,在数量融资中,利用近实时的计算方法来取代昂贵的蒙特卡洛整合。然而,目前这些广泛观念的缺点在于,就计算成本而言,培训一个合适的神经网络很可能令人望而却步。我们在这里通过对深神经网络的培训采用多层次的方法来解决这一缺陷。更准确地说,我们把贝克等人提出的深层次的学习算法与多层次的蒙特卡洛路径模拟概念结合起来。我们的想法是培训几个神经网络网络,每个网络都具有某种近似的质量和计算的复杂性,而培训数据是从所谓的测算师所引入的所谓测算师级数据。我们表明,在某些假设下,培训过程中的差异会通过对深层神经网络进行培训,而我们只能通过将最精确的测算成一个精度的培训水平来降低培训过程。我们只能在计算出一个精度上进行一个精度的培训,因此,因此可以减少一个精度的精度的精度的计算。