Neural controlled differential equations (CDEs) are the continuous-time analogue of recurrent neural networks, as Neural ODEs are to residual networks, and offer a memory-efficient continuous-time way to model functions of potentially irregular time series. Existing methods for computing the forward pass of a Neural CDE involve embedding the incoming time series into path space, often via interpolation, and using evaluations of this path to drive the hidden state. Here, we use rough path theory to extend this formulation. Instead of directly embedding into path space, we instead represent the input signal over small time intervals through its \textit{log-signature}, which are statistics describing how the signal drives a CDE. This is the approach for solving \textit{rough differential equations} (RDEs), and correspondingly we describe our main contribution as the introduction of Neural RDEs. This extension has a purpose: by generalising the Neural CDE approach to a broader class of driving signals, we demonstrate particular advantages for tackling long time series. In this regime, we demonstrate efficacy on problems of length up to 17k observations and observe significant training speed-ups, improvements in model performance, and reduced memory requirements compared to existing approaches.
翻译:神经控制差异方程式(CDE)是反复神经网络的连续时间类比,因为神经代码是剩余网络,并且提供了一种记忆高效的连续时间方法,以模拟潜在不规则时间序列的功能。计算神经 CDE 的前进通过的方法包括将即将到来的时间序列嵌入路径空间,通常是通过内插,并利用对这条路径的评估来驱动隐藏状态。这里,我们使用粗略路径理论来扩展这一表达方式。我们不是直接嵌入路径空间,而是通过它的\ textit{log-signation} 代表输入信号的短暂间隔,这是描述信号如何驱动CDE的统计数据。这是解决\ textit{rough diffraction 公式}(RDEs)的方法,我们相应地将我们的主要贡献描述为引入神经RDEs。这个扩展的目的是:通过将神经光学 CDE 方法推广到更广泛的驱动信号,我们展示了处理长期序列的特殊优势。在这个制度中,我们展示了在长至17k观察和现有培训速度方法上的效率,并观察到现有的改进模式。