The diversity and domain dependence of time series data pose significant challenges in transferring learning to time series forecasting. In this study, we examine the effectiveness of using a transformer model that has been pre-trained on natural language or image data and then fine-tuned for time series forecasting with minimal modifications, specifically, without altering the self-attention and feedforward layers of the residual blocks. This model, known as the Frozen Pretrained Transformer (FPT), is evaluated through fine-tuning on time series forecasting tasks under Zero-Shot, Few-Shot, and normal sample size conditions. Our results demonstrate that pre-training on natural language or images can lead to a comparable or state-of-the-art performance in cross-modality time series forecasting tasks, in contrast to previous studies that focused on fine-tuning within the same modality as the pre-training data. Additionally, we provide a comprehensive theoretical analysis of the universality and the functionality of the FPT. The code is publicly available at https://anonymous.4open.science/r/Pretrained-LM-for-TSForcasting-C561.
翻译:时间序列数据的多样性和领域依赖性在将学习转移到时间序列预测方面提出了重大挑战。在本研究中,我们研究了使用经过自然语言或图像数据培训前的变压器模型的有效性,然后对时间序列的预测进行了微调,并作了微小的修改,具体地说,没有改变剩余区块的自留和进食性向上层。这一模型被称为FFFT, 是通过对Zero-Shot、Phot-Shot和正常样本大小条件下的时间序列预测任务进行微调来加以评价的。我们的结果表明,在跨现代时间序列预测任务中,对自然语言或图像进行预先培训可导致可比的或最先进的性能,而以前的研究的重点是在与培训前数据相同的模式内进行微调。此外,我们对FPT(FPT)的普遍性和功能进行全面的理论分析。代码可在https://anonymous.4open.science/r/prained-LMfor-TSFOR-C561公开查阅。