IEEE Trans Image Process. 2017 Sep;26(9):4321-4330. doi: 10.1109/TIP.2017.2713048. Epub 2017 Jun 7.
Unsupervised outlier detection is a vital task and has high impact on a wide variety of applications domains, such as image analysis and video surveillance. It also gains long-standing attentions and has been extensively studied in multiple research areas. Detecting and taking action on outliers as quickly as possible are imperative in order to protect network and related stakeholders or to maintain the reliability of critical systems. However, outlier detection is difficult due to the one class nature and challenges in feature construction. Sequential anomaly detection is even harder with more challenges from temporal correlation in data, as well as the presence of noise and high dimensionality. In this paper, we introduce a novel deep structured framework to solve the challenging sequential outlier detection problem. We use autoencoder models to capture the intrinsic difference between outliers and normal instances and integrate the models to recurrent neural networks that allow the learning to make use of previous context as well as make the learners more robust to warp along the time axis. Furthermore, we propose to use a layerwise training procedure, which significantly simplifies the training procedure and hence helps achieve efficient and scalable training. In addition, we investigate a fine-tuning step to update all parameters set by incorporating the temporal correlation in the sequence. We further apply our proposed models to conduct systematic experiments on five real-world benchmark data sets. Experimental results demonstrate the effectiveness of our model, compared with other state-of-the-art approaches.
无监督异常检测是一项至关重要的任务,对各种应用领域都有很大的影响,例如图像分析和视频监控。它也引起了长期的关注,并在多个研究领域得到了广泛的研究。为了保护网络和相关利益相关者,或维护关键系统的可靠性,尽快检测和采取异常措施是至关重要的。然而,由于一类性质和特征构建的挑战,异常检测是困难的。由于数据中的时间相关性以及噪声和高维性带来了更多的挑战,序列异常检测甚至更加困难。在本文中,我们引入了一种新颖的深度结构框架来解决具有挑战性的序列异常检测问题。我们使用自动编码器模型来捕获异常值和正常实例之间的内在差异,并将模型集成到递归神经网络中,允许学习者利用先前的上下文,并使学习者对沿时间轴的扭曲更具鲁棒性。此外,我们提出使用逐层训练过程,这大大简化了训练过程,从而有助于实现高效和可扩展的训练。此外,我们研究了一个微调步骤,通过在序列中纳入时间相关性来更新所有参数集。我们进一步将我们提出的模型应用于五个真实基准数据集上进行系统实验。实验结果表明,与其他最先进的方法相比,我们的模型是有效的。