College of Information and Electrical Engineering, China Agricultural University, Beijing, Haidian, 100083, China.
College of Science, China Agricultural University, Beijing, Haidian, 100083, China.
Neural Netw. 2021 Oct;142:457-478. doi: 10.1016/j.neunet.2021.06.028. Epub 2021 Jul 4.
Least squares twin support vector machine (LSTSVM) is an effective and efficient learning algorithm for pattern classification. However, the distance in LSTSVM is measured by squared L-norm metric that may magnify the influence of outliers. In this paper, a novel robust least squares twin support vector machine framework is proposed for binary classification, termed as CL-LSTSVM, which utilizes capped L-norm distance metric to reduce the influence of noise and outliers. The goal of CL-LSTSVM is to minimize the capped L-norm intra-class distance dispersion, and eliminate the influence of outliers during training process, where the value of the metric is controlled by the capped parameter, which can ensure better robustness. The proposed metric includes and extends the traditional metrics by setting appropriate values of p and capped parameter. This strategy not only retains the advantages of LSTSVM, but also improves the robustness in solving a binary classification problem with outliers. However, the nonconvexity of metric makes it difficult to optimize. We design an effective iterative algorithm to solve the CL-LSTSVM. In each iteration, two systems of linear equations are solved. Simultaneously, we present some insightful analyses on the computational complexity and convergence of algorithm. Moreover, we extend the CL-LSTSVM to nonlinear classifier and semi-supervised classification. Experiments are conducted on artificial datasets, UCI benchmark datasets, and image datasets to evaluate our method. Under different noise settings and different evaluation criteria, the experiment results show that the CL-LSTSVM has better robustness than state-of-the-art approaches in most cases, which demonstrates the feasibility and effectiveness of the proposed method.
最小二乘孪生支持向量机 (LSTSVM) 是一种用于模式分类的有效和高效的学习算法。然而,LSTSVM 中的距离是通过平方 L-范数度量来衡量的,这可能会放大异常值的影响。在本文中,我们提出了一种用于二分类的新的鲁棒最小二乘孪生支持向量机框架,称为 CL-LSTSVM,它利用有界 L-范数距离度量来减少噪声和异常值的影响。CL-LSTSVM 的目标是最小化有界 L-范数类内距离分散,在训练过程中消除异常值的影响,其中度量的取值由有界参数控制,可以保证更好的鲁棒性。所提出的度量包括并扩展了传统度量,通过设置适当的 p 值和有界参数值。这种策略不仅保留了 LSTSVM 的优点,而且提高了解决带有异常值的二分类问题的鲁棒性。然而,度量的非凸性使得优化变得困难。我们设计了一种有效的迭代算法来解决 CL-LSTSVM。在每次迭代中,都要求解两个线性方程组。同时,我们对算法的计算复杂度和收敛性进行了一些深入的分析。此外,我们将 CL-LSTSVM 扩展到非线性分类器和半监督分类。在人工数据集、UCI 基准数据集和图像数据集上进行实验来评估我们的方法。在不同的噪声设置和不同的评估标准下,实验结果表明,在大多数情况下,CL-LSTSVM 比最先进的方法具有更好的鲁棒性,这证明了所提出方法的可行性和有效性。