• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

线性递归特征机可证明能恢复低秩矩阵。

Linear Recursive Feature Machines provably recover low-rank matrices.

作者信息

Radhakrishnan Adityanarayanan, Belkin Mikhail, Drusvyatskiy Dmitriy

机构信息

Applied Math, Harvard University, MA 02138.

Broad Institute of Massachusetts Institute of Technology and Harvard, MA 02138.

出版信息

Proc Natl Acad Sci U S A. 2025 Apr;122(13):e2411325122. doi: 10.1073/pnas.2411325122. Epub 2025 Mar 28.

DOI:10.1073/pnas.2411325122
PMID:40153460
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12002225/
Abstract

A fundamental problem in machine learning is to understand how neural networks make accurate predictions, while seemingly bypassing the curse of dimensionality. A possible explanation is that common training algorithms for neural networks implicitly perform dimensionality reduction-a process called feature learning. Recent work [A. Radhakrishnan, D. Beaglehole, P. Pandit, M. Belkin, , 1461-1467 (2024).] posited that the effects of feature learning can be elicited from a classical statistical estimator called the average gradient outer product (AGOP). The authors proposed Recursive Feature Machines (RFMs) as an algorithm that explicitly performs feature learning by alternating between 1) reweighting the feature vectors by the AGOP and 2) learning the prediction function in the transformed space. In this work, we develop theoretical guarantees for how RFM performs dimensionality reduction by focusing on the class of overparameterized problems arising in sparse linear regression and low-rank matrix recovery. Specifically, we show that RFM restricted to linear models (lin-RFM) reduces to a variant of the well-studied Iteratively Reweighted Least Squares (IRLS) algorithm. Furthermore, our results connect feature learning in neural networks and classical sparse recovery algorithms and shed light on how neural networks recover low rank structure from data. In addition, we provide an implementation of lin-RFM that scales to matrices with millions of missing entries. Our implementation is faster than the standard IRLS algorithms since it avoids forming singular value decompositions. It also outperforms deep linear networks for sparse linear regression and low-rank matrix completion.

摘要

机器学习中的一个基本问题是理解神经网络如何做出准确预测,同时似乎避开了维度灾难。一种可能的解释是,神经网络的常见训练算法会隐式地执行降维——这一过程称为特征学习。最近的工作[A. 拉德哈克里什南、D. 比格霍尔、P. 潘迪特、M. 贝尔金, ,1461 - 1467 (2024)]认为,特征学习的效果可以从一种称为平均梯度外积(AGOP)的经典统计估计器中得出。作者提出了递归特征机器(RFM)作为一种算法,该算法通过在1)用AGOP对特征向量重新加权和2)在变换后的空间中学习预测函数之间交替,来显式地执行特征学习。在这项工作中,我们通过关注稀疏线性回归和低秩矩阵恢复中出现的过参数化问题类别,为RFM如何执行降维开发了理论保证。具体来说,我们表明限制在线性模型上的RFM(lin - RFM)简化为经过充分研究的迭代加权最小二乘法(IRLS)算法变体。此外,我们的结果将神经网络中的特征学习与经典稀疏恢复算法联系起来,并阐明了神经网络如何从数据中恢复低秩结构。此外,我们提供了lin - RFM的一种实现,它可以扩展到具有数百万个缺失条目的矩阵。我们的实现比标准的IRLS算法更快,因为它避免了形成奇异值分解。它在稀疏线性回归和低秩矩阵补全方面也优于深度线性网络。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/19eb/12002225/753d2304a569/pnas.2411325122fig04.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/19eb/12002225/aa95cde782dc/pnas.2411325122fig01.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/19eb/12002225/a73ba39aeea5/pnas.2411325122fig02.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/19eb/12002225/35cec0bd5da1/pnas.2411325122fig03.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/19eb/12002225/753d2304a569/pnas.2411325122fig04.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/19eb/12002225/aa95cde782dc/pnas.2411325122fig01.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/19eb/12002225/a73ba39aeea5/pnas.2411325122fig02.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/19eb/12002225/35cec0bd5da1/pnas.2411325122fig03.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/19eb/12002225/753d2304a569/pnas.2411325122fig04.jpg

相似文献

1
Linear Recursive Feature Machines provably recover low-rank matrices.线性递归特征机可证明能恢复低秩矩阵。
Proc Natl Acad Sci U S A. 2025 Apr;122(13):e2411325122. doi: 10.1073/pnas.2411325122. Epub 2025 Mar 28.
2
Mechanism for feature learning in neural networks and backpropagation-free machine learning models.神经网络和无反向传播机器学习模型中的特征学习机制。
Science. 2024 Mar 29;383(6690):1461-1467. doi: 10.1126/science.adi5639. Epub 2024 Mar 7.
3
Sparse subspace clustering for data with missing entries and high-rank matrix completion.用于处理带有缺失值的数据的稀疏子空间聚类及高秩矩阵补全
Neural Netw. 2017 Sep;93:36-44. doi: 10.1016/j.neunet.2017.04.005. Epub 2017 Apr 25.
4
Can a Hebbian-like learning rule be avoiding the curse of dimensionality in sparse distributed data?赫布型学习规则能否避免稀疏分布数据的维度灾难?
Biol Cybern. 2024 Dec;118(5-6):267-276. doi: 10.1007/s00422-024-00995-y. Epub 2024 Sep 9.
5
Dynamics in Deep Classifiers Trained with the Square Loss: Normalization, Low Rank, Neural Collapse, and Generalization Bounds.使用平方损失训练的深度分类器中的动力学:归一化、低秩、神经崩溃和泛化界。
Research (Wash D C). 2023 Mar 8;6:0024. doi: 10.34133/research.0024. eCollection 2023.
6
Sparse low-rank separated representation models for learning from data.用于从数据中学习的稀疏低秩分离表示模型。
Proc Math Phys Eng Sci. 2019 Jan;475(2221):20180490. doi: 10.1098/rspa.2018.0490. Epub 2019 Jan 9.
7
Smoothed low rank and sparse matrix recovery by iteratively reweighted least squares minimization.通过迭代重加权最小二乘法恢复平滑低秩和稀疏矩阵。
IEEE Trans Image Process. 2015 Feb;24(2):646-54. doi: 10.1109/TIP.2014.2380155. Epub 2014 Dec 12.
8
Sparse factorization of square matrices with application to neural attention modeling.方阵的稀疏分解及其在神经注意建模中的应用。
Neural Netw. 2022 Aug;152:160-168. doi: 10.1016/j.neunet.2022.04.014. Epub 2022 Apr 22.
9
Synthetic Lethality Screening with Recursive Feature Machines.使用递归特征机器的合成致死筛选
bioRxiv. 2023 Dec 5:2023.12.03.569803. doi: 10.1101/2023.12.03.569803.
10
CROification: Accurate Kernel Classification with the Efficiency of Sparse Linear SVM.CRO 化:借助稀疏线性支持向量机的效率实现精确内核分类
IEEE Trans Pattern Anal Mach Intell. 2019 Jan;41(1):34-48. doi: 10.1109/TPAMI.2017.2785313. Epub 2017 Dec 19.

本文引用的文献

1
Mechanism for feature learning in neural networks and backpropagation-free machine learning models.神经网络和无反向传播机器学习模型中的特征学习机制。
Science. 2024 Mar 29;383(6690):1461-1467. doi: 10.1126/science.adi5639. Epub 2024 Mar 7.
2
Highly accurate protein structure prediction for the human proteome.高精准度的人类蛋白质组蛋白结构预测。
Nature. 2021 Aug;596(7873):590-596. doi: 10.1038/s41586-021-03828-1. Epub 2021 Jul 22.
3
CVXPY: A Python-Embedded Modeling Language for Convex Optimization.CVXPY:一种用于凸优化的嵌入Python的建模语言。
J Mach Learn Res. 2016 Apr;17.
4
Eigenfaces for recognition.特征脸识别。
J Cogn Neurosci. 1991 Winter;3(1):71-86. doi: 10.1162/jocn.1991.3.1.71.