• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

要理解双重下降,我们需要理解 VC 理论。

To understand double descent, we need to understand VC theory.

机构信息

Department of Electrical and Computer Engineering, University of Minnesota, Twin Cities, Minneapolis, 55455, MN, USA.

Department of Electrical and Computer Engineering, University of Minnesota, Twin Cities, Minneapolis, 55455, MN, USA.

出版信息

Neural Netw. 2024 Jan;169:242-256. doi: 10.1016/j.neunet.2023.10.014. Epub 2023 Oct 16.

DOI:10.1016/j.neunet.2023.10.014
PMID:37913656
Abstract

We analyze generalization performance of over-parameterized learning methods for classification, under VC-theoretical framework. Recently, practitioners in Deep Learning discovered 'double descent' phenomenon, when large networks can fit perfectly available training data, and at the same time, achieve good generalization for future (test) data. The current consensus view is that VC-theoretical results cannot account for good generalization performance of Deep Learning networks. In contrast, this paper shows that double descent can be explained by VC-theoretical concepts, such as VC-dimension and Structural Risk Minimization. We also present empirical results showing that double descent generalization curves can be accurately modeled using classical VC-generalization bounds. Proposed VC-theoretical analysis enables better understanding of generalization curves for data sets with different statistical characteristics, such as low vs high-dimensional data and noisy data. In addition, we analyze generalization performance of transfer learning using pre-trained Deep Learning networks.

摘要

我们在 VC 理论框架下分析了过参数化学习方法在分类方面的泛化性能。最近,深度学习领域的从业者发现了“双下降”现象,即大型网络可以完美地拟合可用的训练数据,同时对未来(测试)数据也能取得良好的泛化效果。目前的共识观点是,VC 理论结果无法解释深度学习网络的良好泛化性能。相比之下,本文表明,双下降现象可以用 VC 理论的概念来解释,如 VC 维数和结构风险最小化。我们还提出了实证结果,表明使用经典的 VC 泛化界可以准确地对双下降泛化曲线进行建模。所提出的 VC 理论分析可以更好地理解具有不同统计特征的数据的泛化曲线,例如低维与高维数据以及噪声数据。此外,我们还分析了使用预训练的深度学习网络的迁移学习的泛化性能。

相似文献

1
To understand double descent, we need to understand VC theory.要理解双重下降,我们需要理解 VC 理论。
Neural Netw. 2024 Jan;169:242-256. doi: 10.1016/j.neunet.2023.10.014. Epub 2023 Oct 16.
2
Understanding Double Descent Using VC-Theoretical Framework.使用VC理论框架理解双重下降
IEEE Trans Neural Netw Learn Syst. 2024 Dec;35(12):18838-18847. doi: 10.1109/TNNLS.2024.3388873. Epub 2024 Dec 2.
3
Model complexity control for regression using VC generalization bounds.使用VC泛化界进行回归的模型复杂度控制
IEEE Trans Neural Netw. 1999;10(5):1075-89. doi: 10.1109/72.788648.
4
Understanding quantum machine learning also requires rethinking generalization.理解量子机器学习还需要重新思考泛化。
Nat Commun. 2024 Mar 13;15(1):2277. doi: 10.1038/s41467-024-45882-z.
5
VC-dimension of univariate decision trees.单变量决策树的 VC 维数。
IEEE Trans Neural Netw Learn Syst. 2015 Feb;26(2):378-87. doi: 10.1109/TNNLS.2014.2385837. Epub 2015 Jan 13.
6
Stability analysis of stochastic gradient descent for homogeneous neural networks and linear classifiers.随机梯度下降在同质神经网络和线性分类器中的稳定性分析。
Neural Netw. 2023 Jul;164:382-394. doi: 10.1016/j.neunet.2023.04.028. Epub 2023 Apr 25.
7
Another look at statistical learning theory and regularization.再探统计学习理论与正则化。
Neural Netw. 2009 Sep;22(7):958-69. doi: 10.1016/j.neunet.2009.04.005. Epub 2009 Apr 22.
8
The Vapnik-Chervonenkis dimension of graph and recursive neural networks.图和递归神经网络的 Vapnik-Chervonenkis 维数。
Neural Netw. 2018 Dec;108:248-259. doi: 10.1016/j.neunet.2018.08.010. Epub 2018 Sep 1.
9
An Optimal Transport Analysis on Generalization in Deep Learning.深度学习中的泛化的最优传输分析。
IEEE Trans Neural Netw Learn Syst. 2023 Jun;34(6):2842-2853. doi: 10.1109/TNNLS.2021.3109942. Epub 2023 Jun 1.
10
Improving generalization of deep neural networks by leveraging margin distribution.利用边缘分布提高深度神经网络的泛化能力。
Neural Netw. 2022 Jul;151:48-60. doi: 10.1016/j.neunet.2022.03.019. Epub 2022 Mar 17.