• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

相似文献

1
Information bottleneck theory of high-dimensional regression: relevancy, efficiency and optimality.高维回归的信息瓶颈理论:相关性、效率与最优性
Adv Neural Inf Process Syst. 2022 Dec;35:9784-9796.
2
Reconciling modern machine-learning practice and the classical bias-variance trade-off.调和现代机器学习实践与经典偏差-方差权衡。
Proc Natl Acad Sci U S A. 2019 Aug 6;116(32):15849-15854. doi: 10.1073/pnas.1903070116. Epub 2019 Jul 24.
3
A decision-theoretic approach to the evaluation of machine learning algorithms in computational drug discovery.一种基于决策理论的计算药物发现中机器学习算法评估方法。
Bioinformatics. 2019 Nov 1;35(22):4656-4663. doi: 10.1093/bioinformatics/btz293.
4
On the Information Bottleneck Problems: Models, Connections, Applications and Information Theoretic Views.关于信息瓶颈问题:模型、联系、应用及信息论观点
Entropy (Basel). 2020 Jan 27;22(2):151. doi: 10.3390/e22020151.
5
Exact and Soft Successive Refinement of the Information Bottleneck.信息瓶颈的精确与软渐进细化
Entropy (Basel). 2023 Sep 19;25(9):1355. doi: 10.3390/e25091355.
6
Neurodynamics-driven holistic approaches to semi-supervised feature selection.基于神经动力学的半监督特征选择整体方法。
Neural Netw. 2023 Jan;157:377-386. doi: 10.1016/j.neunet.2022.10.029. Epub 2022 Nov 3.
7
Memorizing without overfitting: Bias, variance, and interpolation in overparameterized models.记忆而不过度拟合:超参数化模型中的偏差、方差和插值
Phys Rev Res. 2022 Mar-May;4(1). doi: 10.1103/physrevresearch.4.013201. Epub 2022 Mar 15.
8
Bias-variance decomposition of overparameterized regression with random linear features.具有随机线性特征的过参数化回归的偏差-方差分解
Phys Rev E. 2022 Aug;106(2-2):025304. doi: 10.1103/PhysRevE.106.025304.
9
SURPRISES IN HIGH-DIMENSIONAL RIDGELESS LEAST SQUARES INTERPOLATION.高维无脊最小二乘插值中的意外情况。
Ann Stat. 2022 Apr;50(2):949-986. doi: 10.1214/21-aos2133. Epub 2022 Apr 7.
10
The Deterministic Information Bottleneck.确定性信息瓶颈
Neural Comput. 2017 Jun;29(6):1611-1630. doi: 10.1162/NECO_a_00961. Epub 2017 Apr 14.

引用本文的文献

1
Nonlinear classification of neural manifolds with contextual information.具有上下文信息的神经流形的非线性分类。
Phys Rev E. 2025 Mar;111(3-2):035302. doi: 10.1103/PhysRevE.111.035302.
2
Exploring the Trade-Off in the Variational Information Bottleneck for Regression with a Single Training Run.单次训练运行下探索回归变分信息瓶颈中的权衡
Entropy (Basel). 2024 Nov 30;26(12):1043. doi: 10.3390/e26121043.
3
Extrinsic vs Intrinsic Criticality in Systems with Many Components.具有多个组件的系统中的外在与内在关键性
ArXiv. 2023 Sep 25:arXiv:2309.13898v1.

本文引用的文献

1
Perturbation Theory for the Information Bottleneck.信息瓶颈的微扰理论
Adv Neural Inf Process Syst. 2021 Dec;34:21008-21018.
2
SURPRISES IN HIGH-DIMENSIONAL RIDGELESS LEAST SQUARES INTERPOLATION.高维无脊最小二乘插值中的意外情况。
Ann Stat. 2022 Apr;50(2):949-986. doi: 10.1214/21-aos2133. Epub 2022 Apr 7.
3
Gaussian Information Bottleneck and the Non-Perturbative Renormalization Group.高斯信息瓶颈与非微扰重整化群
New J Phys. 2022 Mar;24(3). doi: 10.1088/1367-2630/ac395d. Epub 2022 Mar 9.
4
Trading bits in the readout from a genetic network.从遗传网络的读出中交易位。
Proc Natl Acad Sci U S A. 2021 Nov 16;118(46). doi: 10.1073/pnas.2109011118.
5
Relevance in the Renormalization Group and in Information Theory.重整化群与信息论中的相关性。
Phys Rev Lett. 2021 Jun 18;126(24):240601. doi: 10.1103/PhysRevLett.126.240601.
6
Maximally efficient prediction in the early fly visual system may support evasive flight maneuvers.在早期果蝇视觉系统中实现最高效的预测,可能有助于其做出逃避飞行的机动动作。
PLoS Comput Biol. 2021 May 20;17(5):e1008965. doi: 10.1371/journal.pcbi.1008965. eCollection 2021 May.
7
Benign overfitting in linear regression.线性回归中的良性过拟合。
Proc Natl Acad Sci U S A. 2020 Dec 1;117(48):30063-30070. doi: 10.1073/pnas.1907378117. Epub 2020 Apr 24.
8
Reconciling modern machine-learning practice and the classical bias-variance trade-off.调和现代机器学习实践与经典偏差-方差权衡。
Proc Natl Acad Sci U S A. 2019 Aug 6;116(32):15849-15854. doi: 10.1073/pnas.1903070116. Epub 2019 Jul 24.
9
The Information Bottleneck and Geometric Clustering.信息瓶颈与几何聚类
Neural Comput. 2019 Mar;31(3):596-612. doi: 10.1162/neco_a_01136. Epub 2018 Oct 12.
10
Information Dropout: Learning Optimal Representations Through Noisy Computation.信息丢失:通过噪声计算学习最优表示
IEEE Trans Pattern Anal Mach Intell. 2018 Dec;40(12):2897-2905. doi: 10.1109/TPAMI.2017.2784440. Epub 2018 Jan 10.

高维回归的信息瓶颈理论:相关性、效率与最优性

Information bottleneck theory of high-dimensional regression: relevancy, efficiency and optimality.

作者信息

Ngampruetikorn Vudtiwat, Schwab David J

机构信息

Initiative for the Theoretical Sciences, The Graduate Center, CUNY.

出版信息

Adv Neural Inf Process Syst. 2022 Dec;35:9784-9796.

PMID:37332888
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10275337/
Abstract

Avoiding overfitting is a central challenge in machine learning, yet many large neural networks readily achieve zero training loss. This puzzling contradiction necessitates new approaches to the study of overfitting. Here we quantify overfitting via residual information, defined as the bits in fitted models that encode noise in training data. Information efficient learning algorithms minimize residual information while maximizing the relevant bits, which are predictive of the unknown generative models. We solve this optimization to obtain the information content of optimal algorithms for a linear regression problem and compare it to that of randomized ridge regression. Our results demonstrate the fundamental trade-off between residual and relevant information and characterize the relative information efficiency of randomized regression with respect to optimal algorithms. Finally, using results from random matrix theory, we reveal the information complexity of learning a linear map in high dimensions and unveil information-theoretic analogs of double and multiple descent phenomena.

摘要

避免过拟合是机器学习中的核心挑战,然而许多大型神经网络很容易实现零训练损失。这种令人困惑的矛盾需要新的方法来研究过拟合。在这里,我们通过残差信息来量化过拟合,残差信息定义为拟合模型中编码训练数据噪声的比特。信息高效学习算法在最大化相关比特的同时最小化残差信息,这些相关比特可预测未知的生成模型。我们求解此优化问题以获得线性回归问题的最优算法的信息内容,并将其与随机岭回归的信息内容进行比较。我们的结果展示了残差信息与相关信息之间的基本权衡,并刻画了随机回归相对于最优算法的相对信息效率。最后,利用随机矩阵理论的结果,我们揭示了高维中学习线性映射的信息复杂性,并揭示了双重和多重下降现象的信息论类似物。