Time series forecasting is a key component in many industrial and business decision processes and recurrent neural network (RNN) based models have achieved impressive progress on various time series forecasting tasks. However, most of the existing methods focus on single-task forecasting problems by learning separately based on limited supervised objectives, which often suffer from insufficient training instances. As the Transformer architecture and other attention-based models have demonstrated its great capability of capturing long term dependency, we propose two self-attention based sharing schemes for multi-task time series forecasting which can train jointly across multiple tasks. We augment a sequence of paralleled Transformer encoders with an external public multi-head attention function, which is updated by all data of all tasks. Experiments on a number of real-world multi-task time series forecasting tasks show that our proposed architectures can not only outperform the state-of-the-art single-task forecasting baselines but also outperform the RNN-based multi-task forecasting method.
翻译:时间序列预测是许多工业和商业决策过程的一个关键组成部分,基于经常神经网络的模型在各种时间序列预测任务方面取得了令人印象深刻的进展,然而,大多数现有方法侧重于单任务预测问题,根据有限的监督目标分别学习,而这些目标往往缺乏足够的培训。由于变换器结构和其他关注模型显示了它捕捉长期依赖性的巨大能力,我们提议了两种基于自我注意的多任务时间序列预测共享计划,可以对多个任务进行联合培训。我们增加了一系列平行的变换器编码器与外部公共多头关注功能的序列,由所有任务的数据加以更新。关于一些现实世界多任务时间序列预测任务的实验表明,我们拟议的结构不仅能够超越最先进的单一任务预测基线,而且能够超越基于RNN的多任务预测方法。