The conjoining of dynamical systems and deep learning has become a topic of great interest. In particular, neural differential equations (NDEs) demonstrate that neural networks and differential equation are two sides of the same coin. Traditional parameterised differential equations are a special case. Many popular neural network architectures, such as residual networks and recurrent networks, are discretisations. NDEs are suitable for tackling generative problems, dynamical systems, and time series (particularly in physics, finance, ...) and are thus of interest to both modern machine learning and traditional mathematical modelling. NDEs offer high-capacity function approximation, strong priors on model space, the ability to handle irregular data, memory efficiency, and a wealth of available theory on both sides. This doctoral thesis provides an in-depth survey of the field. Topics include: neural ordinary differential equations (e.g. for hybrid neural/mechanistic modelling of physical systems); neural controlled differential equations (e.g. for learning functions of irregular time series); and neural stochastic differential equations (e.g. to produce generative models capable of representing complex stochastic dynamics, or sampling from complex high-dimensional distributions). Further topics include: numerical methods for NDEs (e.g. reversible differential equations solvers, backpropagation through differential equations, Brownian reconstruction); symbolic regression for dynamical systems (e.g. via regularised evolution); and deep implicit models (e.g. deep equilibrium models, differentiable optimisation). We anticipate this thesis will be of interest to anyone interested in the marriage of deep learning with dynamical systems, and hope it will provide a useful reference for the current state of the art.
翻译:动态系统和深层学习的结合已成为引起极大兴趣的话题。 特别是, 神经差异方程式( NDEs) 表明神经网络和差异方程式是同一硬币的两面。 传统的参数化差异方程式是一个特例。 许多流行的神经网络结构, 如残余网络和经常性网络, 是分解的。 NDEs 适合解决基因化问题、动态系统和时间序列( 特别是在物理、 财务、... ), 因而对现代机器学习和传统的深层数学建模都感兴趣。 NDEs 提供了高能力功能的近距离、 模型空间的强前端、 处理不正常数据的能力、 记忆效率以及两侧的丰富可用理论。 这个博士式网络结构提供了对实地的深入调查。 主题包括: 神经普通差异方程式( 例如, 用于物理系统的混合神经/机械建模); 神经控制差异参考方程式( 例如, 用于学习不规则化的时间序列; 以及 神经分析差异方程式的精确度变异方方程式( e. g. realtocal checkal diversal dal deal deal deality sality sality sqal sqal) diqal sqal) (egraqal) aqal) aqaldalisqalislation.