We present a representation learning framework for financial time series forecasting. One challenge of using deep learning models for finance forecasting is the shortage of available training data when using small datasets. Direct trend classification using deep neural networks trained on small datasets is susceptible to the overfitting problem. In this paper, we propose to first learn compact representations from time series data, then use the learned representations to train a simpler model for predicting time series movements. We consider a class-conditioned latent variable model. We train an encoder network to maximize the mutual information between the latent variables and the trend information conditioned on the encoded observed variables. We show that conditional mutual information maximization can be approximated by a contrastive loss. Then, the problem is transformed into a classification task of determining whether two encoded representations are sampled from the same class or not. This is equivalent to performing pairwise comparisons of the training datapoints, and thus, improves the generalization ability of the encoder network. We use deep autoregressive models as our encoder to capture long-term dependencies of the sequence data. Empirical experiments indicate that our proposed method has the potential to advance state-of-the-art performance.
翻译:我们为财务时间序列预测提供了一个代表性学习框架。在使用小型数据集时,利用深学习模型进行财务预测的挑战之一是缺乏现有培训数据。使用在小数据集方面受过训练的深神经网络进行直接趋势分类很容易造成问题。在本文中,我们提议首先从时间序列数据中学习缩略图,然后利用所学的表述方法来训练一个更简单的模型来预测时间序列变化。我们考虑一个有等级条件的潜在潜在变数模型。我们训练了一个编码网络,以尽量扩大潜在变数与以编码观测到的变量为条件的趋势信息之间的相互信息。我们显示,有条件的相互信息最大化可以被对比性损失所近似。然后,问题变成一个分类任务,即确定两个编码的表达方式是否从同一类别中抽样,这相当于对培训数据点进行对齐比较,从而提高编码网络的总体能力。我们使用深自导模型作为我们的编码,用以捕捉测测测测测测测序数据的长期依赖性。Enpiral实验表明,我们提出的方法具有前进状态的潜力。