Neural networks inspired by differential equations have proliferated for the past several years. Neural ordinary differential equations (NODEs) and neural controlled differential equations (NCDEs) are two representative examples of them. In theory, NCDEs provide better representation learning capability for time-series data than NODEs. In particular, it is known that NCDEs are suitable for processing irregular time-series data. Whereas NODEs have been successfully extended after adopting attention, however, it had not been studied yet how to integrate attention into NCDEs. To this end, we present the method of Attentive Neural Controlled Differential Equations (ANCDEs) for time-series classification and forecasting, where dual NCDEs are used: one for generating attention values, and the other for evolving hidden vectors for a downstream machine learning task. We conduct experiments with three real-world time-series datasets and 10 baselines. After dropping some values, we also conduct irregular time-series experiments. Our method consistently shows the best accuracy in all cases by non-trivial margins. Our visualizations also show that the presented attention mechanism works as intended by focusing on crucial information.
翻译:在过去几年中,受差异方程式启发的神经网络已经扩散了。神经普通差异方程式和神经控制差异方程式是其中两个具有代表性的例子。理论上,NCDEs为时间序列数据提供了比NODEs更好的代表性学习能力。特别是,众所周知,NCDEs适合处理非正常时间序列数据。虽然NDEs在被关注之后成功地扩展了,但是,没有研究如何将注意力纳入NCDEs。为此,我们提出了用于时间序列分类和预测的Attenal Control drominal Equations(ANDEs)方法,其中使用了双重NCDEs:一个用于引起注意值,另一个用于为下游机学习任务开发隐藏矢量。我们用三个真实世界的时间序列数据集和10个基线进行实验。在降低某些值之后,我们还进行了不规则的时间序列实验。我们的方法始终以非三边边边为对象,显示所有案例的最佳准确性。我们所展示的注意机制也显示,以关键信息为重点。