Zhao Haifeng, Wang Siqi, Wang Zheng
College of Computer Science and Technology, Anhui University, Hefei 230601, China
Center for OPTical IMagery Analysis and Learning, Northwestern Polytechnical University, Xi'an 710072, China
Neural Comput. 2018 Oct;30(10):2781-2804. doi: 10.1162/neco_a_01116. Epub 2018 Jul 18.
Least squares regression (LSR) is a fundamental statistical analysis technique that has been widely applied to feature learning. However, limited by its simplicity, the local structure of data is easy to neglect, and many methods have considered using orthogonal constraint for preserving more local information. Another major drawback of LSR is that the loss function between soft regression results and hard target values cannot precisely reflect the classification ability; thus, the idea of the large margin constraint is put forward. As a consequence, we pay attention to the concepts of large margin and orthogonal constraint to propose a novel algorithm, orthogonal least squares regression with large margin (OLSLM), for multiclass classification in this letter. The core task of this algorithm is to learn regression targets from data and an orthogonal transformation matrix simultaneously such that the proposed model not only ensures every data point can be correctly classified with a large margin than conventional least squares regression, but also can preserve more local data structure information in the subspace. Our efficient optimization method for solving the large margin constraint and orthogonal constraint iteratively proved to be convergent in both theory and practice. We also apply the large margin constraint in the process of generating a sparse learning model for feature selection via joint [Formula: see text]-norm minimization on both loss function and regularization terms. Experimental results validate that our method performs better than state-of-the-art methods on various real-world data sets.
最小二乘回归(LSR)是一种基本的统计分析技术,已广泛应用于特征学习。然而,由于其简单性的限制,数据的局部结构很容易被忽略,许多方法已经考虑使用正交约束来保留更多的局部信息。LSR的另一个主要缺点是软回归结果与硬目标值之间的损失函数不能精确反映分类能力;因此,提出了大间隔约束的思想。因此,在本文中,我们关注大间隔和正交约束的概念,提出了一种用于多类分类的新算法——大间隔正交最小二乘回归(OLSLM)。该算法的核心任务是同时从数据和正交变换矩阵中学习回归目标,使得所提出的模型不仅能确保每个数据点都能以比传统最小二乘回归更大的间隔被正确分类,而且能在子空间中保留更多的局部数据结构信息。我们用于迭代求解大间隔约束和正交约束的高效优化方法在理论和实践中都被证明是收敛的。我们还在通过对损失函数和正则化项进行联合[公式:见原文]范数最小化来生成用于特征选择的稀疏学习模型的过程中应用大间隔约束。实验结果验证了我们的方法在各种真实世界数据集上比现有方法表现更好。