Li Benhan, Zhang Wei, Lu Mingxin
School of Information Management, Nanjing University, Nanjing, 210023, China.
School of Computer Science, Nanjing University of Posts and Telecommunications, Nanjing, 210023, China.
Neural Netw. 2025 Aug;188:107484. doi: 10.1016/j.neunet.2025.107484. Epub 2025 Apr 23.
Recently, Transformer-based and multilayer perceptron (MLP) based architectures have formed a competitive landscape in the field of time series forecasting. There is evidence that series decomposition can further enhance the model's ability to perceive temporal patterns. However, most of the existing Transformer-based decomposed models capture seasonal features progressively and assist in adding trends for forecasting, but ignore the deep information contained in trends and may lead to pattern mismatch in the fusion stage. In addition, the permutation invariance of the attention mechanism inevitably leads to the loss of temporal order. After in-depth analysis of the applicability of attention and linear layers to series components, we propose to use attention to learn multivariate correlations from trends, and MLP to capture seasonal patterns. We further introduce an integrated codec that provides the same multivariate relationship representation for both the encoding and decoding stages, ensuring effective inheritance of temporal dependencies. To mitigate the fading of sequentiality during attention, we propose trend enhancement module, which maintains the stability of the trend by expanding the series to a longer time scale, helping the attention mechanism to achieve fine-grained feature representations. Extensive experiments show that our model exhibits state-of-the-art prediction performance on large-scale datasets.
最近,基于Transformer和基于多层感知器(MLP)的架构在时间序列预测领域形成了竞争态势。有证据表明,序列分解可以进一步增强模型感知时间模式的能力。然而,大多数现有的基于Transformer的分解模型逐步捕捉季节性特征并辅助添加趋势进行预测,但忽略了趋势中包含的深层信息,可能导致融合阶段的模式不匹配。此外,注意力机制的排列不变性不可避免地导致时间顺序的丢失。在深入分析注意力和线性层对序列组件的适用性后,我们建议使用注意力从趋势中学习多变量相关性,并使用MLP捕捉季节性模式。我们进一步引入了一种集成编解码器,它为编码和解码阶段提供相同的多变量关系表示,确保时间依赖性的有效继承。为了减轻注意力过程中序列性的衰减,我们提出了趋势增强模块,通过将序列扩展到更长的时间尺度来保持趋势的稳定性,帮助注意力机制实现细粒度特征表示。大量实验表明,我们的模型在大规模数据集上表现出了领先的预测性能。