• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

一种基于广义 -范数回归的特征选择算法。

A generalized -norm regression based feature selection algorithm.

作者信息

Zhi X, Liu J, Wu S, Niu C

机构信息

School of Science, Xi'an University of Posts and Telecommunications, Xi'an, People's Republic of China.

School of Communication and Information Engineering, Xi'an University of Posts and Telecommunications, Xi'an, People's Republic of China.

出版信息

J Appl Stat. 2021 Sep 17;50(3):703-723. doi: 10.1080/02664763.2021.1975662. eCollection 2023.

DOI:10.1080/02664763.2021.1975662
PMID:36819074
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9930865/
Abstract

Feature selection is an important data dimension reduction method, and it has been used widely in applications involving high-dimensional data such as genetic data analysis and image processing. In order to achieve robust feature selection, the latest works apply the or -norm of matrix to the loss function and regularization terms in regression, and have achieved encouraging results. However, these existing works rigidly set the matrix norms used in the loss function and the regularization terms to the same or -norm, which limit their applications. In addition, the algorithms for solutions they present either have high computational complexity and are not suitable for large data sets, or cannot provide satisfying performance due to the approximate calculation. To address these problems, we present a generalized -norm regression based feature selection ( -RFS) method based on a new optimization criterion. The criterion extends the optimization criterion of ( -RFS) when the loss function and the regularization terms in regression use different matrix norms. We cast the new optimization criterion in a regression framework without regularization. In this framework, the new optimization criterion can be solved using an iterative re-weighted least squares (IRLS) procedure in which the least squares problem can be solved efficiently by using the least square QR decomposition (LSQR) algorithm. We have conducted extensive experiments to evaluate the proposed algorithm on various well-known data sets of both gene expression and image data sets, and compare it with other related feature selection methods.

摘要

特征选择是一种重要的数据降维方法,已广泛应用于涉及高维数据的应用中,如基因数据分析和图像处理。为了实现稳健的特征选择,最新的研究工作将矩阵的 或 -范数应用于回归中的损失函数和正则化项,并取得了令人鼓舞的成果。然而,这些现有工作将损失函数和正则化项中使用的矩阵范数严格设置为相同的 或 -范数,这限制了它们的应用。此外,他们提出的求解算法要么计算复杂度高,不适用于大数据集,要么由于近似计算而无法提供令人满意的性能。为了解决这些问题,我们基于一种新的优化准则提出了一种基于广义 -范数回归的特征选择( -RFS)方法。当回归中的损失函数和正则化项使用不同的矩阵范数时,该准则扩展了( -RFS)的优化准则。我们将新的优化准则置于无正则化的回归框架中。在这个框架中,新的优化准则可以使用迭代加权最小二乘法(IRLS)来求解,其中最小二乘问题可以通过使用最小二乘QR分解(LSQR)算法有效地求解。我们进行了广泛的实验,在各种著名的基因表达数据集和图像数据集上评估所提出的算法,并将其与其他相关的特征选择方法进行比较。

相似文献

1
A generalized -norm regression based feature selection algorithm.一种基于广义 -范数回归的特征选择算法。
J Appl Stat. 2021 Sep 17;50(3):703-723. doi: 10.1080/02664763.2021.1975662. eCollection 2023.
2
L2-norm multiple kernel learning and its application to biomedical data fusion.L2-范数多核学习及其在生物医学数据融合中的应用。
BMC Bioinformatics. 2010 Jun 8;11:309. doi: 10.1186/1471-2105-11-309.
3
Low-dose CT reconstruction via L1 dictionary learning regularization using iteratively reweighted least-squares.基于迭代重加权最小二乘法的 L1 字典学习正则化的低剂量 CT 重建
Biomed Eng Online. 2016 Jun 18;15(1):66. doi: 10.1186/s12938-016-0193-y.
4
Noise suppression for dual-energy CT via penalized weighted least-square optimization with similarity-based regularization.基于相似性正则化的惩罚加权最小二乘优化用于双能CT的噪声抑制
Med Phys. 2016 May;43(5):2676. doi: 10.1118/1.4947485.
5
Unsupervised Feature Selection With Constrained ℓ₂,₀-Norm and Optimized Graph.基于约束ℓ₂,₀范数和优化图的无监督特征选择
IEEE Trans Neural Netw Learn Syst. 2022 Apr;33(4):1702-1713. doi: 10.1109/TNNLS.2020.3043362. Epub 2022 Apr 4.
6
A Fast and Accurate Matrix Completion Method Based on QR Decomposition and L -Norm Minimization.一种基于QR分解和L范数最小化的快速准确矩阵填充方法。
IEEE Trans Neural Netw Learn Syst. 2019 Mar;30(3):803-817. doi: 10.1109/TNNLS.2018.2851957. Epub 2018 Jul 26.
7
Sparse PCA via l-Norm Regularization for Unsupervised Feature Selection.基于 l-范数正则化的稀疏主成分分析用于无监督特征选择
IEEE Trans Pattern Anal Mach Intell. 2023 Apr;45(4):5322-5328. doi: 10.1109/TPAMI.2021.3121329. Epub 2023 Mar 7.
8
Generalized and Robust Least Squares Regression.
IEEE Trans Neural Netw Learn Syst. 2024 May;35(5):7006-7020. doi: 10.1109/TNNLS.2022.3213594. Epub 2024 May 2.
9
Joint embedding learning and sparse regression: a framework for unsupervised feature selection.联合嵌入学习和稀疏回归:一种无监督特征选择的框架。
IEEE Trans Cybern. 2014 Jun;44(6):793-804. doi: 10.1109/TCYB.2013.2272642. Epub 2013 Jul 22.
10
Unsupervised Feature Selection With Flexible Optimal Graph.基于灵活最优图的无监督特征选择
IEEE Trans Neural Netw Learn Syst. 2024 Feb;35(2):2014-2027. doi: 10.1109/TNNLS.2022.3186171. Epub 2024 Feb 5.

引用本文的文献

1
Editorial to the special issue: Statistical Approaches for Big Data and Machine Learning.特刊社论:大数据与机器学习的统计方法
J Appl Stat. 2023 Feb 7;50(3):451-455. doi: 10.1080/02664763.2023.2162471. eCollection 2023.

本文引用的文献

1
2D Feature Selection by Sparse Matrix Regression.基于稀疏矩阵回归的二维特征选择。
IEEE Trans Image Process. 2017 Sep;26(9):4255-4268. doi: 10.1109/TIP.2017.2713948. Epub 2017 Jun 8.
2
Effective Discriminative Feature Selection With Nontrivial Solution.有效鉴别特征选择与非平凡解。
IEEE Trans Neural Netw Learn Syst. 2016 Apr;27(4):796-808. doi: 10.1109/TNNLS.2015.2424721. Epub 2015 May 14.
3
Algorithm 937: MINRES-QLP for Symmetric and Hermitian Linear Equations and Least-Squares Problems.算法937:用于对称和埃尔米特线性方程及最小二乘问题的MINRES-QLP算法
ACM Trans Math Softw. 2014 Feb;40(2). doi: 10.1145/2527267.
4
Local coordinates alignment with global preservation for dimensionality reduction.局部坐标与全局保形对齐进行降维。
IEEE Trans Neural Netw Learn Syst. 2013 Jan;24(1):106-17. doi: 10.1109/TNNLS.2012.2225844.
5
Feature selection based on mutual information: criteria of max-dependency, max-relevance, and min-redundancy.基于互信息的特征选择:最大依赖、最大相关和最小冗余准则。
IEEE Trans Pattern Anal Mach Intell. 2005 Aug;27(8):1226-38. doi: 10.1109/TPAMI.2005.159.
6
Minimum redundancy feature selection from microarray gene expression data.从微阵列基因表达数据中进行最小冗余特征选择。
J Bioinform Comput Biol. 2005 Apr;3(2):185-205. doi: 10.1142/s0219720005001004.