Chang Zhuoqing, Liu Shubo, Qiu Run, Song Song, Cai Zhaohui, Tu Guoqing
School of Computer Science, Wuhan University, 299# Bayi Rd, Wuchang District, Wuhan, 430072 Hubei China.
School of Cyber Science and Engineering, Wuhan University, 299# Bayi Rd, Wuchang District, Wuhan, 430072 Hubei China.
J Supercomput. 2023 May 18:1-29. doi: 10.1007/s11227-023-05327-8.
Internet of Things realizes the ubiquitous connection of all things, generating countless time-tagged data called time series. However, real-world time series are often plagued with missing values on account of noise or malfunctioning sensors. Existing methods for modeling such incomplete time series typically involve preprocessing steps, such as deletion or missing data imputation using statistical learning or machine learning methods. Unfortunately, these methods unavoidable destroy time information and bring error accumulation to the subsequent model. To this end, this paper introduces a novel continuous neural network architecture, named Time-aware Neural-Ordinary Differential Equations (TN-ODE), for incomplete time data modeling. The proposed method not only supports imputation missing values at arbitrary time points, but also enables multi-step prediction at desired time points. Specifically, TN-ODE employs a time-aware Long Short-Term Memory as an encoder, which effectively learns the posterior distribution from partial observed data. Additionally, the derivative of latent states is parameterized with a fully connected network, thereby enabling continuous-time latent dynamics generation. The proposed TN-ODE model is evaluated on both real-world and synthetic incomplete time-series datasets by conducting data interpolation and extrapolation tasks as well as classification task. Extensive experiments show the TN-ODE model outperforms baseline methods in terms of Mean Square Error for imputation and prediction tasks, as well as accuracy in downstream classification task.
物联网实现了万物的普遍连接,产生了无数带有时间标签的数据,即时间序列。然而,由于噪声或传感器故障,现实世界中的时间序列常常存在缺失值。现有的对这种不完整时间序列进行建模的方法通常涉及预处理步骤,例如使用统计学习或机器学习方法进行删除或缺失数据插补。不幸的是,这些方法不可避免地会破坏时间信息,并给后续模型带来误差累积。为此,本文引入了一种新颖的连续神经网络架构,名为时间感知神经常微分方程(TN-ODE),用于不完整时间数据建模。所提出的方法不仅支持在任意时间点插补缺失值,还能在期望的时间点进行多步预测。具体而言,TN-ODE采用时间感知长短期记忆作为编码器,它能有效地从部分观测数据中学习后验分布。此外,潜在状态的导数由全连接网络进行参数化,从而实现连续时间潜在动态生成。通过进行数据插值和外推任务以及分类任务,在所提出的TN-ODE模型在真实世界和合成不完整时间序列数据集上进行了评估。大量实验表明,TN-ODE模型在插补和预测任务的均方误差方面以及下游分类任务的准确率方面均优于基线方法。