Multivariate time series forecasting (MTSF) is a fundamental problem in numerous real-world applications. Recently, Transformer has become the de facto solution for MTSF, especially for the long-term cases. However, except for the one forward operation, the basic configurations in existing MTSF Transformer architectures were barely carefully verified. In this study, we point out that the current tokenization strategy in MTSF Transformer architectures ignores the token uniformity inductive bias of Transformers. Therefore, the vanilla MTSF transformer struggles to capture details in time series and presents inferior performance. Based on this observation, we make a series of evolution on the basic architecture of the vanilla MTSF transformer. We vary the flawed tokenization strategy, along with the decoder structure and embeddings. Surprisingly, the evolved simple transformer architecture is highly effective, which successfully avoids the over-smoothing phenomena in the vanilla MTSF transformer, achieves a more detailed and accurate prediction, and even substantially outperforms the state-of-the-art Transformers that are well-designed for MTSF.
翻译:多变时间序列预测(MTSF)是许多现实应用中的一个基本问题。 最近,变异器已成为MTF的事实上解决方案,特别是长期的。 但是,除了一个前方操作外,现有的MTSF变异器结构的基本配置几乎没有经过仔细核实。 在本研究中,我们指出,MTSF变异器结构中目前的代谢化战略忽视了变异器的象征性统一感化偏向。因此,香草花变异器在时间序列中捕捉细节并显示低效性能。基于这一观察,我们在香草变异器的基本结构上进行了一系列演进。我们把有缺陷的代号化战略与解码器结构和嵌入相去。 令人惊讶的是,进化的简单变异器结构非常有效,成功地避免了香草变异器的过度移动现象,实现了更加详细和准确的预测,甚至大大超越了为MTF设计完善的状态型变异器。