Time Series Foundation Models (TSFMs) have shown significant impact through their model capacity, scalability, and zero-shot generalization. However, due to the heterogeneity of inter-variate dependencies and the backbone scalability on large-scale multivariate datasets, most TSFMs are typically pre-trained on univariate time series. This limitation renders them oblivious to crucial information from diverse covariates in real-world forecasting tasks. To further enhance the performance of TSFMs, we propose a general covariate-aware adaptation (CoRA) framework for TSFMs. It leverages pre-trained backbones of foundation models while effectively incorporating exogenous covariates from various modalities, including time series, language, and images, to improve the quality of predictions. Technically, CoRA maintains the equivalence of initialization and parameter consistency during adaptation. With preserved backbones of foundation models as frozen feature extractors, the outcome embeddings from foundation models are empirically demonstrated more informative than raw data. Further, CoRA employs a novel Granger Causality Embedding (GCE) to automatically evaluate covariates regarding their causal predictability with respect to the target variate. We incorporate these weighted embeddings with a zero-initialized condition-injection mechanism, avoiding catastrophic forgetting of pre-trained foundation models and gradually integrates exogenous information. Extensive experiments show that CoRA of TSFMs surpasses state-of-the-art covariate-aware deep forecasters with full or few-shot training samples, achieving 31.1% MSE reduction on covariate-aware forecasting. Compared to other adaptation methods, CoRA exhibits strong compatibility with various advanced TSFMs and extends the scope of covariates to other modalities, presenting a practical paradigm for the application of TSFMs.
翻译:时间序列基础模型(TSFMs)凭借其模型容量、可扩展性和零样本泛化能力展现出显著影响力。然而,由于变量间依赖关系的异质性以及大规模多元数据集上的主干网络可扩展性问题,大多数TSFM通常基于单变量时间序列进行预训练。这一局限使其无法感知现实世界预测任务中来自多样化协变量的关键信息。为进⼀步提升TSFM的性能,我们提出了⼀种通⽤的协变量感知自适应(CoRA)框架。该框架利⽤基础模型的预训练主干网络,同时有效整合来自不同模态(包括时间序列、语⾔和图像)的外⽣协变量,以提升预测质量。技术上,CoRA在自适应过程中保持初始化等价性与参数⼀致性。通过将基础模型的保留主干作为冻结特征提取器,实验证明基础模型输出的嵌入表示⽐原始数据更具信息量。此外,CoRA采⽤新颖的格兰杰因果嵌入(GCE)⾃动评估协变量相对于⽬标变量的因果预测能⼒。我们通过零初始化条件注⼊机制整合这些加权嵌入,避免了预训练基础模型的灾难性遗忘,并逐步融合外⽣信息。⼤量实验表明,基于TSFM的CoRA在使⽤全量或少样本训练数据时均优于最先进的协变量感知深度预测模型,在协变量感知预测任务上实现了31.1%的均⽅误差降低。与其他⾃适应⽅法相⽐,CoRA展现出与各类先进TSFM的强兼容性,并将协变量范围扩展⾄其他模态,为TSFM的实际应⽤提供了实⽤范式。