Time series forecasting is at the core of important application domains posing significant challenges to machine learning algorithms. Recently neural network architectures have been widely applied to the problem of time series forecasting. Most of these models are trained by minimizing a loss function that measures predictions' deviation from the real values. Typical loss functions include mean squared error (MSE) and mean absolute error (MAE). In the presence of noise and uncertainty, neural network models tend to replicate the last observed value of the time series, thus limiting their applicability to real-world data. In this paper, we provide a formal definition of the above problem and we also give some examples of forecasts where the problem is observed. We also propose a regularization term penalizing the replication of previously seen values. We evaluate the proposed regularization term both on synthetic and real-world datasets. Our results indicate that the regularization term mitigates to some extent the aforementioned problem and gives rise to more robust models.
翻译:时间序列预测是重要应用领域的核心,对机器学习算法构成重大挑战。最近神经网络结构被广泛应用于时间序列预测问题。这些模型大多通过最大限度地减少损失函数来培训,以衡量预测偏离实际价值的情况。典型损失函数包括平均平方错误(MSE)和平均绝对错误(MAE)。在出现噪音和不确定性的情况下,神经网络模型往往复制时间序列最后观察到的价值,从而限制其对现实世界数据的可适用性。本文中,我们提供了上述问题的正式定义,我们还提供了一些观察到问题的预测例子。我们还提出了一个规范化术语,惩罚复制以前所看到的价值。我们评估了合成和现实世界数据集的拟议正规化术语。我们的结果显示,正规化术语在一定程度上缓解了上述问题,并产生了更强有力的模型。