Faculty of Mathematics and Computer Science, Transilvania University of Braşov, Braşov, Romania.
Computer Science Department, Central Washington University, Ellensburg, WA, USA; Transilvania University of Braşov, Braşov, Romania.
Neural Netw. 2022 Aug;152:528-541. doi: 10.1016/j.neunet.2022.05.018. Epub 2022 May 27.
We study the problem of learning the data samples' distribution as it changes in time. This change, known as concept drift, complicates the task of training a model, as the predictions become less and less accurate. It is known that Support Vector Machines (SVMs) can learn weighted input instances and that they can also be trained online (incremental-decremental learning). Combining these two SVM properties, the open problem is to define an online SVM concept drift model with shifting weighted window. The classic SVM model should be retrained from scratch after each window shift. We introduce the Weighted Incremental-Decremental SVM (WIDSVM), a generalization of the incremental-decremental SVM for shifting windows. WIDSVM is capable of learning from data streams with concept drift, using the weighted shifting window technique. The soft margin constrained optimization problem imposed on the shifting window is reduced to an incremental-decremental SVM. At each window shift, we determine the exact conditions for vector migration during the incremental-decremental process. We perform experiments on artificial and real-world concept drift datasets; they show that the classification accuracy of WIDSVM significantly improves compared to a SVM with no shifting window. The WIDSVM training phase is fast, since it does not retrain from scratch after each window shift.
我们研究了数据样本随时间变化的分布学习问题。这种变化,即概念漂移,使得训练模型的任务变得更加复杂,因为预测的准确性会越来越低。已知支持向量机(SVM)可以学习加权输入实例,并且它们也可以在线(增量-递减学习)进行训练。结合这两个 SVM 属性,开放式问题是定义具有移位加权窗口的在线 SVM 概念漂移模型。经典的 SVM 模型应该在每次窗口移位后从头开始重新训练。我们引入了加权增量-递减 SVM(WIDSVM),这是一种用于移位窗口的增量-递减 SVM 的推广。WIDSVM 能够使用加权移位窗口技术从具有概念漂移的数据流中学习。施加在移位窗口上的软间隔约束优化问题被简化为增量-递减 SVM。在每次窗口移位时,我们确定在增量-递减过程中向量迁移的确切条件。我们在人工和真实世界的概念漂移数据集上进行实验,结果表明 WIDSVM 的分类准确性与没有移位窗口的 SVM 相比有显著提高。WIDSVM 的训练阶段很快,因为它不需要在每次窗口移位后从头开始重新训练。