Zhang Xiangyue, Li Chao, Ji Ling, Kang Yuyun, Pan Mingming, Liu Zhuo, Qi Qiang
School of Information Science and Engineering, Linyi University, Linyi, 276000, China.
Daopuyun (Shandong) Intelligent Technology Co., Ltd, Jinan, 265200, China.
Sci Rep. 2025 Jul 29;15(1):27668. doi: 10.1038/s41598-025-11375-2.
Traffic forecasting is considered a cornerstone of smart city development. A key challenge is capturing the long-term spatiotemporal dependencies of traffic data while improving the model's generalization ability. To address these issues, various sophisticated modules are embedded into different models. However, this approach increases the computational cost of the model. Additionally, adding or replacing datasets in a trained model requires retraining, which decreases prediction accuracy and increases time cost. To address the challenges faced by existing models in handling long-term spatiotemporal dependencies and high computational costs, this study proposes an enhanced pre-training method called the Improved Spatiotemporal Diffusion Graph (ImPreSTDG). While existing traffic prediction models, particularly those based on Graph Convolutional Networks (GCNs) and deep learning, are effective at capturing short-term spatiotemporal dependencies, they often experience accuracy degradation and increased computational demands when dealing with long-term dependencies. To overcome these limitations, we introduce a Denoised Diffusion Probability Model (DDPM) as part of the pre-training process, which enhances the model's ability to learn from long-term spatiotemporal data while significantly reducing computational costs. During the pre-training phase, ImPreSTDG employs a data masking and recovery strategy, with DDPM facilitating the reconstruction of masked data segments, thereby enabling the model to capture long-term dependencies in the traffic data. Additionally, we propose the Mamba module, which leverages the Selective State Space Model (SSM) to effectively capture long-term multivariate spatiotemporal correlations. This module enables more efficient processing of long sequences, extracting essential patterns while minimizing computational resource consumption. By improving computational efficiency, the Mamba module addresses the challenge of modeling long-term dependencies without compromising accuracy in capturing extended spatiotemporal trends. In the fine-tuning phase, the decoder is replaced with a forecasting header, and the pre-trained parameters are frozen. The forecasting header includes a meta-learning fusion module and a spatiotemporal convolutional layer, which facilitates the integration of both long-term and short-term traffic data for accurate forecasting. The model is then trained and adapted to the specific forecasting task. Experiments conducted on three real-world traffic datasets demonstrate that the proposed pre-training method significantly enhances the model's ability to handle long-term dependencies, missing data, and high computational costs, providing a more efficient solution for traffic prediction.
交通流量预测被视为智慧城市发展的基石。一个关键挑战是捕捉交通数据的长期时空依赖性,同时提高模型的泛化能力。为了解决这些问题,各种复杂的模块被嵌入到不同的模型中。然而,这种方法增加了模型的计算成本。此外,在训练好的模型中添加或替换数据集需要重新训练,这会降低预测准确性并增加时间成本。为了应对现有模型在处理长期时空依赖性和高计算成本方面面临的挑战,本研究提出了一种增强的预训练方法,称为改进的时空扩散图(ImPreSTDG)。虽然现有的交通预测模型,特别是基于图卷积网络(GCN)和深度学习的模型,在捕捉短期时空依赖性方面很有效,但在处理长期依赖性时,它们往往会出现准确性下降和计算需求增加的情况。为了克服这些限制,我们引入了一个去噪扩散概率模型(DDPM)作为预训练过程的一部分,它增强了模型从长期时空数据中学习的能力,同时显著降低了计算成本。在预训练阶段,ImPreSTDG采用数据掩码和恢复策略,DDPM有助于重建掩码数据段,从而使模型能够捕捉交通数据中的长期依赖性。此外,我们提出了曼巴模块,它利用选择性状态空间模型(SSM)有效地捕捉长期多变量时空相关性。该模块能够更高效地处理长序列,提取基本模式,同时最小化计算资源消耗。通过提高计算效率,曼巴模块解决了在不影响捕捉扩展时空趋势准确性的情况下对长期依赖性进行建模的挑战。在微调阶段,解码器被替换为预测头,预训练参数被冻结。预测头包括一个元学习融合模块和一个时空卷积层,它有助于整合长期和短期交通数据以进行准确预测。然后对模型进行训练并使其适应特定的预测任务。在三个真实世界交通数据集上进行的实验表明,所提出的预训练方法显著增强了模型处理长期依赖性、缺失数据和高计算成本的能力,为交通预测提供了更有效的解决方案。