Deep learning model (primarily convolutional networks and LSTM) for time series classification has been studied broadly by the community with the wide applications in different domains like healthcare, finance, industrial engineering and IoT. Meanwhile, Transformer Networks recently achieved frontier performance on various natural language processing and computer vision tasks. In this work, we explored a simple extension of the current Transformer Networks with gating, named Gated Transformer Networks (GTN) for the multivariate time series classification problem. With the gating that merges two towers of Transformer which model the channel-wise and step-wise correlations respectively, we show how GTN is naturally and effectively suitable for the multivariate time series classification task. We conduct comprehensive experiments on thirteen dataset with full ablation study. Our results show that GTN is able to achieve competing results with current state-of-the-art deep learning models. We also explored the attention map for the natural interpretability of GTN on time series modeling. Our preliminary results provide a strong baseline for the Transformer Networks on multivariate time series classification task and grounds the foundation for future research.
翻译:社区广泛研究了时间序列分类的深学习模型(主要是革命网络和LSTM),在保健、金融、工业工程和IoT等不同领域应用了广泛的应用。 同时,变形网络最近在各种自然语言处理和计算机视觉任务方面实现了前沿性能。在这项工作中,我们探索了当前变形网络的简单扩展,包括了名为GGNT(GTN)的多变时间序列分类问题。在将两个变形塔合为一体时序关系模型时,我们展示了GTN如何自然和有效地适合多变式时间序列分类任务。我们进行了13个数据集的全面实验,并进行了全面减缩研究。我们的结果显示,GTN能够与当前最先进的深层学习模型取得相互竞争的结果。我们还探讨了GTN在时间序列模型上自然解释GTN的注意地图。我们的初步结果为变形网络在多变式时间序列分类任务上提供了强有力的基准,并为未来研究奠定了基础。