• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用无偏性换取方差缩减:核回归中用于稳健模型选择的正则化子空间信息准则

Trading variance reduction with unbiasedness: the regularized subspace information criterion for robust model selection in kernel regression.

作者信息

Sugiyama Masashi, Kawanabe Motoaki, Müller Klaus-Robert

机构信息

Fraunhofer FIRST, IDA, 12489 Berlin, Germany.

出版信息

Neural Comput. 2004 May;16(5):1077-104. doi: 10.1162/089976604773135113.

DOI:10.1162/089976604773135113
PMID:15070511
Abstract

A well-known result by Stein (1956) shows that in particular situations, biased estimators can yield better parameter estimates than their generally preferred unbiased counterparts. This letter follows the same spirit, as we will stabilize the unbiased generalization error estimates by regularization and finally obtain more robust model selection criteria for learning. We trade a small bias against a larger variance reduction, which has the beneficial effect of being more precise on a single training set. We focus on the subspace information criterion (SIC), which is an unbiased estimator of the expected generalization error measured by the reproducing kernel Hilbert space norm. SIC can be applied to the kernel regression, and it was shown in earlier experiments that a small regularization of SIC has a stabilization effect. However, it remained open how to appropriately determine the degree of regularization in SIC. In this article, we derive an unbiased estimator of the expected squared error, between SIC and the expected generalization error and propose determining the degree of regularization of SIC such that the estimator of the expected squared error is minimized. Computer simulations with artificial and real data sets illustrate that the proposed method works effectively for improving the precision of SIC, especially in the high-noise-level cases. We furthermore compare the proposed method to the original SIC, the cross-validation, and an empirical Bayesian method in ridge parameter selection, with good results.

摘要

斯坦因(1956年)的一个著名结果表明,在特定情况下,有偏估计量可能会比通常更受青睐的无偏估计量产生更好的参数估计。这封信秉承了同样的精神,因为我们将通过正则化来稳定无偏泛化误差估计,最终获得更稳健的学习模型选择标准。我们用一个小偏差换取更大的方差减少,这在单个训练集上具有更精确的有益效果。我们关注子空间信息准则(SIC),它是由再生核希尔伯特空间范数衡量的期望泛化误差的无偏估计量。SIC可应用于核回归,早期实验表明对SIC进行小的正则化具有稳定作用。然而,如何在SIC中适当地确定正则化程度仍然是一个未解决的问题。在本文中,我们推导了SIC与期望泛化误差之间期望平方误差的无偏估计量,并提出确定SIC的正则化程度,以使期望平方误差的估计量最小化。使用人工和真实数据集的计算机模拟表明,所提出的方法有效地提高了SIC的精度,特别是在高噪声水平的情况下。我们还在岭参数选择中将所提出的方法与原始SIC、交叉验证和经验贝叶斯方法进行了比较,结果良好。

相似文献

1
Trading variance reduction with unbiasedness: the regularized subspace information criterion for robust model selection in kernel regression.用无偏性换取方差缩减:核回归中用于稳健模型选择的正则化子空间信息准则
Neural Comput. 2004 May;16(5):1077-104. doi: 10.1162/089976604773135113.
2
Subspace information criterion for model selection.用于模型选择的子空间信息准则。
Neural Comput. 2001 Aug;13(8):1863-89. doi: 10.1162/08997660152469387.
3
Robust regularized kernel regression.稳健正则化核回归
IEEE Trans Syst Man Cybern B Cybern. 2008 Dec;38(6):1639-44. doi: 10.1109/TSMCB.2008.927279.
4
On the problem in model selection of neural network regression in overrealizable scenario.关于超可实现场景下神经网络回归模型选择中的问题。
Neural Comput. 2002 Aug;14(8):1979-2002. doi: 10.1162/089976602760128090.
5
Gene selection in cancer classification using sparse logistic regression with Bayesian regularization.使用带贝叶斯正则化的稀疏逻辑回归进行癌症分类中的基因选择。
Bioinformatics. 2006 Oct 1;22(19):2348-55. doi: 10.1093/bioinformatics/btl386. Epub 2006 Jul 14.
6
Constructing Bayesian formulations of sparse kernel learning methods.构建稀疏核学习方法的贝叶斯公式。
Neural Netw. 2005 Jun-Jul;18(5-6):674-83. doi: 10.1016/j.neunet.2005.06.002.
7
A comparative investigation on subspace dimension determination.子空间维度确定的比较研究。
Neural Netw. 2004 Oct-Nov;17(8-9):1051-9. doi: 10.1016/j.neunet.2004.07.005.
8
Practical selection of SVM parameters and noise estimation for SVM regression.支持向量机回归中支持向量机参数的实际选择与噪声估计
Neural Netw. 2004 Jan;17(1):113-26. doi: 10.1016/S0893-6080(03)00169-2.
9
Equations of states in singular statistical estimation.奇异统计估计中的状态方程。
Neural Netw. 2010 Jan;23(1):20-34. doi: 10.1016/j.neunet.2009.08.002. Epub 2009 Aug 15.
10
Optimal design of regularization term and regularization parameter by subspace information criterion.基于子空间信息准则的正则化项和正则化参数的优化设计
Neural Netw. 2002 Apr;15(3):349-61. doi: 10.1016/s0893-6080(02)00022-9.