The problem of processing very long time-series data (e.g., a length of more than 10,000) is a long-standing research problem in machine learning. Recently, one breakthrough, called neural rough differential equations (NRDEs), has been proposed and has shown that it is able to process such data. Their main concept is to use the log-signature transform, which is known to be more efficient than the Fourier transform for irregular long time-series, to convert a very long time-series sample into a relatively shorter series of feature vectors. However, the log-signature transform causes non-trivial spatial overheads. To this end, we present the method of LOweR-Dimensional embedding of log-signature (LORD), where we define an NRDE-based autoencoder to implant the higher-depth log-signature knowledge into the lower-depth log-signature. We show that the encoder successfully combines the higher-depth and the lower-depth log-signature knowledge, which greatly stabilizes the training process and increases the model accuracy. In our experiments with benchmark datasets, the improvement ratio by our method is up to 75\% in terms of various classification and forecasting evaluation metrics.
翻译:处理非常长的时间序列数据的问题(例如,10 000多个时间序列的长度)是机器学习中长期存在的研究问题。最近,提出了一项突破,称为神经粗差方程(NRDEs),并表明它能够处理这些数据。它们的主要概念是使用日志签名转换,据知这比非正常的长期时间序列的Fourier转换效率更高,将一个非常长的时间序列样本转换成一个较短的特性矢量序列。然而,日志签名转换导致非三重空间间接间接费用。为此,我们提出了LOWER-DID(LORRR-D)方法,我们在该方法中定义了基于NRDE的自动编码,将更深入的日志签名知识植入低深度的日志签名。我们显示,编码成功地将高深度和低深度的日志签名知识结合在一起,大大稳定了培训进程,提高了模型的准确性。为此,我们用基准数据集进行实验时,我们用各种方法进行的预测和指标分类的改进比率达到75。