IEEE Trans Neural Netw Learn Syst. 2018 Oct;29(10):4967-4982. doi: 10.1109/TNNLS.2017.2785403. Epub 2018 Jan 15.
Feature selection aims to select a subset of features from high-dimensional data according to a predefined selecting criterion. Sparse learning has been proven to be a powerful technique in feature selection. Sparse regularizer, as a key component of sparse learning, has been studied for several years. Although convex regularizers have been used in many works, there are some cases where nonconvex regularizers outperform convex regularizers. To make the process of selecting relevant features more effective, we propose a novel nonconvex sparse metric on matrices as the sparsity regularization in this paper. The new nonconvex regularizer could be written as the difference of the $\ell _{2,1}$ norm and the Frobenius ( $\ell _{2,2}$ ) norm, which is named the $\ell _{2,1-2}$ . To find the solution of the resulting nonconvex formula, we design an iterative algorithm in the framework of ConCave-Convex Procedure (CCCP) and prove its strong global convergence. An adopted alternating direction method of multipliers is embedded to solve the sequence of convex subproblems in CCCP efficiently. Using the scaled cluster indictors of data points as pseudolabels, we also apply $\ell _{2,1-2}$ to the unsupervised case. To the best of our knowledge, it is the first work considering nonconvex regularization for matrices in the unsupervised learning scenario. Numerical experiments are performed on real-world data sets to demonstrate the effectiveness of the proposed method.
特征选择旨在根据预定义的选择标准从高维数据中选择特征子集。稀疏学习已被证明是特征选择中的一种强大技术。稀疏正则化器作为稀疏学习的关键组成部分,已经研究了数年。尽管凸正则化器已在许多工作中使用,但在某些情况下非凸正则化器的性能优于凸正则化器。为了使相关特征的选择过程更有效,本文提出了一种新颖的基于矩阵的非凸稀疏度量作为稀疏正则化。新的非凸正则化器可以写成$\ell_{2,1}$范数与Frobenius($\ell_{2,2}$)范数之差,称为$\ell_{2,1 - 2}$。为了找到所得非凸公式的解,我们在凹凸过程(CCCP)框架内设计了一种迭代算法,并证明了其强全局收敛性。嵌入了一种采用的交替方向乘子法来有效求解CCCP中的凸子问题序列。使用数据点的缩放聚类指标作为伪标签,我们还将$\ell_{2,1 - 2}$应用于无监督情况。据我们所知,这是在无监督学习场景中考虑矩阵非凸正则化的第一项工作。在真实数据集上进行了数值实验,以证明所提方法的有效性。