Straat Michiel, Abadi Fthi, Göpfert Christina, Hammer Barbara, Biehl Michael
Bernoulli Institute for Mathematics, Computer Science and Artificial Intelligence, University of Groningen, Nijenborgh 9, 9747 AG Groningen, The Netherlands.
Center of Excellence-Cognitive Interaction Technology (CITEC), Bielefeld University, Inspiration 1, 33619 Bielefeld, Germany.
Entropy (Basel). 2018 Oct 10;20(10):775. doi: 10.3390/e20100775.
We introduce a modeling framework for the investigation of on-line machine learning processes in non-stationary environments. We exemplify the approach in terms of two specific model situations: In the first, we consider the learning of a classification scheme from clustered data by means of prototype-based Learning Vector Quantization (LVQ). In the second, we study the training of layered neural networks with sigmoidal activations for the purpose of regression. In both cases, the target, i.e., the classification or regression scheme, is considered to change continuously while the system is trained from a stream of labeled data. We extend and apply methods borrowed from statistical physics which have been used frequently for the exact description of training dynamics in stationary environments. Extensions of the approach allow for the computation of typical learning curves in the presence of concept drift in a variety of model situations. First results are presented and discussed for stochastic drift processes in classification and regression problems. They indicate that LVQ is capable of tracking a classification scheme under drift to a non-trivial extent. Furthermore, we show that concept drift can cause the persistence of sub-optimal plateau states in gradient based training of layered neural networks for regression.
我们介绍了一个用于研究非平稳环境中在线机器学习过程的建模框架。我们通过两种特定的模型情况来举例说明该方法:第一种情况,我们考虑通过基于原型的学习向量量化(LVQ)从聚类数据中学习分类方案。第二种情况,我们研究具有 sigmoid 激活函数的分层神经网络的训练,用于回归目的。在这两种情况下,目标,即分类或回归方案,在系统从标记数据流中进行训练时被认为是不断变化的。我们扩展并应用了从统计物理学中借鉴的方法,这些方法经常用于精确描述平稳环境中的训练动态。该方法的扩展允许在各种模型情况下计算存在概念漂移时的典型学习曲线。针对分类和回归问题中的随机漂移过程给出并讨论了初步结果。结果表明,LVQ 能够在一定程度上跟踪漂移下的分类方案。此外,我们表明概念漂移会导致基于梯度的分层神经网络回归训练中次优平稳状态的持续存在。