Laboratoire de Modélisation et Sûreté des Systèmes, Institut Charles Delaunay (UMR CNRS 6279), Universitè de Technologie de Troyes, 12 rue Marie Curie, BP 2060, 10010 Troyes cedex, France.
IEEE Trans Pattern Anal Mach Intell. 2012 Sep;34(9):1814-26. doi: 10.1109/TPAMI.2011.270.
Kernel principal component analysis (kernel-PCA) is an elegant nonlinear extension of one of the most used data analysis and dimensionality reduction techniques, the principal component analysis. In this paper, we propose an online algorithm for kernel-PCA. To this end, we examine a kernel-based version of Oja's rule, initially put forward to extract a linear principal axe. As with most kernel-based machines, the model order equals the number of available observations. To provide an online scheme, we propose to control the model order. We discuss theoretical results, such as an upper bound on the error of approximating the principal functions with the reduced-order model. We derive a recursive algorithm to discover the first principal axis, and extend it to multiple axes. Experimental results demonstrate the effectiveness of the proposed approach, both on synthetic data set and on images of handwritten digits, with comparison to classical kernel-PCA and iterative kernel-PCA.
核主成分分析(kernel-PCA)是最常用的数据分析和降维技术之一的主成分分析的一种优雅的非线性扩展。在本文中,我们提出了一种核主成分分析的在线算法。为此,我们研究了基于核的 Oja 规则的版本,该规则最初用于提取线性主轴。与大多数基于核的机器一样,模型阶数等于可用观测值的数量。为了提供在线方案,我们建议控制模型阶数。我们讨论了理论结果,例如用降阶模型近似主函数的误差的上界。我们推导出一种递归算法来发现第一个主轴,并将其扩展到多个轴。实验结果表明,该方法在合成数据集和手写数字图像上都具有有效性,与经典核主成分分析和迭代核主成分分析进行了比较。