Cawley Gavin C, Talbot Nicola L C
School of Computing Sciences, University of East Anglia, Norwich NR4 7TJ, UK.
Neural Netw. 2004 Dec;17(10):1467-75. doi: 10.1016/j.neunet.2004.07.002.
Leave-one-out cross-validation has been shown to give an almost unbiased estimator of the generalisation properties of statistical models, and therefore provides a sensible criterion for model selection and comparison. In this paper we show that exact leave-one-out cross-validation of sparse Least-Squares Support Vector Machines (LS-SVMs) can be implemented with a computational complexity of only O(ln2) floating point operations, rather than the O(l2n2) operations of a naïve implementation, where l is the number of training patterns and n is the number of basis vectors. As a result, leave-one-out cross-validation becomes a practical proposition for model selection in large scale applications. For clarity the exposition concentrates on sparse least-squares support vector machines in the context of non-linear regression, but is equally applicable in a pattern recognition setting.
留一法交叉验证已被证明能给出统计模型泛化特性的几乎无偏估计,因此为模型选择和比较提供了一个合理的标准。在本文中,我们表明,稀疏最小二乘支持向量机(LS-SVM)的精确留一法交叉验证可以以仅O(ln2)浮点运算的计算复杂度来实现,而不是朴素实现中的O(l2n2)运算,其中l是训练模式的数量,n是基向量的数量。因此,留一法交叉验证成为大规模应用中模型选择的一个实际方案。为清晰起见,阐述集中在非线性回归背景下的稀疏最小二乘支持向量机,但同样适用于模式识别设置。