Time series data is ubiquitous in research as well as in a wide variety of industrial applications. Effectively analyzing the available historical data and providing insights into the far future allows us to make effective decisions. Recent research has witnessed the superior performance of transformer-based architectures, especially in the regime of far horizon time series forecasting. However, the current state of the art sparse Transformer architectures fail to couple down- and upsampling procedures to produce outputs in a similar resolution as the input. We propose the Yformer model, based on a novel Y-shaped encoder-decoder architecture that (1) uses direct connection from the downscaled encoder layer to the corresponding upsampled decoder layer in a U-Net inspired architecture, (2) Combines the downscaling/upsampling with sparse attention to capture long-range effects, and (3) stabilizes the encoder-decoder stacks with the addition of an auxiliary reconstruction loss. Extensive experiments have been conducted with relevant baselines on four benchmark datasets, demonstrating an average improvement of 19.82, 18.41 percentage MSE and 13.62, 11.85 percentage MAE in comparison to the current state of the art for the univariate and the multivariate settings respectively.
翻译:在研究和各种工业应用中,时间序列数据都是普遍存在的。有效分析现有历史数据并洞察远方的未来,使我们能够做出有效决定。最近的研究见证了以变压器为基础的结构的优异性表现,特别是在远地时间序列预测制度下。然而,目前最稀疏的变压器结构的状况未能将下层和上层取样程序相匹配,以产生与输入相类似的分辨率的产出。我们提议了Y型新颖的编码器脱coder结构,该结构(1) 利用从下层编码器层与U-Net激励型结构中相应高采样的解码层直接连接,(2) 将降幅/加插图与很少注意捕捉远程效应结合起来,(3) 稳定编码器脱钩器堆,加上辅助性重建损失。我们用四个基准数据集的相关基线进行了广泛的实验,表明,从19.82、18.41 %的MSE和13.62、11.85 %的MAE多变异性分别与当前艺术的状态进行比较。