Deep neural networks, when optimized with sufficient data, provide accurate representations of high-dimensional functions; in contrast, function approximation techniques that have predominated in scientific computing do not scale well with dimensionality. As a result, many high-dimensional sampling and approximation problems once thought intractable are being revisited through the lens of machine learning. While the promise of unparalleled accuracy may suggest a renaissance for applications that require parameterizing representations of complex systems, in many applications gathering sufficient data to develop such a representation remains a significant challenge. Here we introduce an approach that combines rare events sampling techniques with neural network optimization to optimize objective functions that are dominated by rare events. We show that importance sampling reduces the asymptotic variance of the solution to a learning problem, suggesting benefits for generalization. We study our algorithm in the context of learning dynamical transition pathways between two states of a system, a problem with applications in statistical physics and implications in machine learning theory. Our numerical experiments demonstrate that we can successfully learn even with the compounding difficulties of high-dimension and rare data.
翻译:深神经网络,如果以足够数据优化,则能准确表达高维功能;相反,在科学计算中占主导地位的功能近似技术与维度相比并不大。因此,许多高维抽样和近似问题一旦被认为难以解决,就会通过机器学习的透镜重新审视。 虽然无与伦比的准确性承诺可能意味着需要将复杂系统的描述参数化的应用的复兴,但在许多应用中,收集足够数据以发展这种表达法仍然是一个重大挑战。在这里,我们采用了一种方法,将稀有事件取样技术和神经网络优化结合起来,以优化由稀有事件主导的客观功能。我们表明,重要取样可以减少解决学习问题的方法的无足轻重差异,为普遍化带来好处。我们是在学习系统两个状态之间的动态过渡路径的背景下研究我们的算法,这是统计物理应用和机器学习理论所涉问题的一个问题。我们的数字实验表明,即使高维度和稀有数据的复杂困难也能够成功学习。