One of the major characteristics of financial time series is that they contain a large amount of non-stationary noise, which is challenging for deep neural networks. People normally use various features to address this problem. However, the performance of these features depends on the choice of hyper-parameters. In this paper, we propose to use neural networks to represent these indicators and train a large network constructed of smaller networks as feature layers to fine-tune the prior knowledge represented by the indicators. During back propagation, prior knowledge is transferred from human logic to machine logic via gradient descent. Prior knowledge is the deep belief of neural network and teaches the network to not be affected by non-stationary noise. Moreover, co-distillation is applied to distill the structure into a much smaller size to reduce redundant features and the risk of overfitting. In addition, the decisions of the smaller networks in terms of gradient descent are more robust and cautious than those of large networks. In numerical experiments, we find that our algorithm is faster and more accurate than traditional methods on real financial datasets. We also conduct experiments to verify and comprehend the method.
翻译:金融时序的主要特征之一是,它们含有大量的非静止噪音,这对深层神经网络来说具有挑战性。人们通常使用各种特征来解决这个问题。但是,这些特征的性能取决于对超参数的选择。在本文件中,我们提议使用神经网络来代表这些指标,并训练由较小网络组成的大型网络,作为特征层来微调指标所代表的先前的知识。在后期传播过程中,先前的知识通过梯度下降从人类逻辑向机器逻辑转移。先前的知识是神经网络的深刻信念,并教导网络不受非静止噪音的影响。此外,共同蒸馏还用于将结构蒸馏成一个小得多的体积,以减少冗余特征和过大的风险。此外,较小网络在梯度下降方面的决定比大网络更有力、更谨慎。在数字实验中,我们发现我们的算法比实际财务数据集的传统方法更快、更准确。我们还进行实验,以核实和理解该方法。