• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

核密度估计、核方法和大数据集的快速学习。

Kernel Density Estimation, Kernel Methods, and Fast Learning in Large Data Sets.

出版信息

IEEE Trans Cybern. 2014 Jan;44(1):1-20. doi: 10.1109/TSMCB.2012.2236828. Epub 2013 Jun 18.

DOI:10.1109/TSMCB.2012.2236828
PMID:23797315
Abstract

Kernel methods such as the standard support vector machine and support vector regression trainings take O(N(3)) time and O(N(2)) space complexities in their naïve implementations, where N is the training set size. It is thus computationally infeasible in applying them to large data sets, and a replacement of the naive method for finding the quadratic programming (QP) solutions is highly desirable. By observing that many kernel methods can be linked up with kernel density estimate (KDE) which can be efficiently implemented by some approximation techniques, a new learning method called fast KDE (FastKDE) is proposed to scale up kernel methods. It is based on establishing a connection between KDE and the QP problems formulated for kernel methods using an entropy-based integrated-squared-error criterion. As a result, FastKDE approximation methods can be applied to solve these QP problems. In this paper, the latest advance in fast data reduction via KDE is exploited. With just a simple sampling strategy, the resulted FastKDE method can be used to scale up various kernel methods with a theoretical guarantee that their performance does not degrade a lot. It has a time complexity of O(m(3)) where m is the number of the data points sampled from the training set. Experiments on different benchmarking data sets demonstrate that the proposed method has comparable performance with the state-of-art method and it is effective for a wide range of kernel methods to achieve fast learning in large data sets.

摘要

核方法,如标准支持向量机和支持向量回归训练,在其朴素实现中,时间复杂度为 O(N(3)),空间复杂度为 O(N(2)),其中 N 是训练集的大小。因此,将其应用于大数据集是不切实际的,非常需要一种替代朴素方法来寻找二次规划 (QP) 解的方法。通过观察到许多核方法可以与核密度估计 (KDE) 联系起来,而 KDE 可以通过一些近似技术有效地实现,因此提出了一种称为快速 KDE (FastKDE) 的新学习方法来扩展核方法。它基于在基于熵的积分均方误差准则下,在 KDE 和为核方法制定的 QP 问题之间建立联系。结果,可以应用 FastKDE 近似方法来解决这些 QP 问题。在本文中,利用了通过 KDE 进行快速数据缩减的最新进展。通过简单的采样策略,所得到的 FastKDE 方法可以用于扩展各种核方法,理论上保证其性能不会大幅下降。它的时间复杂度为 O(m(3)),其中 m 是从训练集中采样的数据点的数量。在不同基准数据集上的实验表明,该方法具有与最先进方法相当的性能,并且对于广泛的核方法在大数据集中实现快速学习是有效的。

相似文献

1
Kernel Density Estimation, Kernel Methods, and Fast Learning in Large Data Sets.核密度估计、核方法和大数据集的快速学习。
IEEE Trans Cybern. 2014 Jan;44(1):1-20. doi: 10.1109/TSMCB.2012.2236828. Epub 2013 Jun 18.
2
Generalized core vector machines.广义核向量机
IEEE Trans Neural Netw. 2006 Sep;17(5):1126-40. doi: 10.1109/TNN.2006.878123.
3
L2 kernel classification.L2 核分类。
IEEE Trans Pattern Anal Mach Intell. 2010 Oct;32(10):1822-31. doi: 10.1109/TPAMI.2009.188.
4
The maximum vector-angular margin classifier and its fast training on large datasets using a core vector machine.最大向量角边缘分类器及其在大数据集上使用核心向量机的快速训练。
Neural Netw. 2012 Mar;27:60-73. doi: 10.1016/j.neunet.2011.10.005. Epub 2011 Oct 21.
5
Clustered Nyström method for large scale manifold learning and dimension reduction.用于大规模流形学习和降维的聚类奈斯特罗姆方法。
IEEE Trans Neural Netw. 2010 Oct;21(10):1576-87. doi: 10.1109/TNN.2010.2064786. Epub 2010 Aug 30.
6
Ranking Support Vector Machine with Kernel Approximation.带核近似的排序支持向量机
Comput Intell Neurosci. 2017;2017:4629534. doi: 10.1155/2017/4629534. Epub 2017 Feb 13.
7
Stochastic subset selection for learning with kernel machines.用于核机器学习的随机子集选择
IEEE Trans Syst Man Cybern B Cybern. 2012 Jun;42(3):616-26. doi: 10.1109/TSMCB.2011.2171680. Epub 2011 Oct 27.
8
Multiscale asymmetric orthogonal wavelet kernel for linear programming support vector learning and nonlinear dynamic systems identification.多尺度非对称正交小波核的线性规划支持向量学习和非线性动态系统辨识。
IEEE Trans Cybern. 2014 May;44(5):712-24. doi: 10.1109/TCYB.2013.2279834. Epub 2013 Sep 18.
9
Localized Multiple Kernel Learning Via Sample-Wise Alternating Optimization.基于样本交替优化的局部多核学习
IEEE Trans Cybern. 2014 Jan;44(1):137-48. doi: 10.1109/TCYB.2013.2248710. Epub 2013 Mar 22.
10
Direct Kernel Perceptron (DKP): ultra-fast kernel ELM-based classification with non-iterative closed-form weight calculation.直接核感知机(DKP):基于超快速核极限学习机的分类方法,具有非迭代的闭式权重计算。
Neural Netw. 2014 Feb;50:60-71. doi: 10.1016/j.neunet.2013.11.002. Epub 2013 Nov 14.

引用本文的文献

1
A predictive model for therapy failure in patients with chronic myeloid leukemia receiving tyrosine kinase inhibitor therapy.慢性髓性白血病患者接受酪氨酸激酶抑制剂治疗后治疗失败的预测模型。
Blood. 2024 Oct 31;144(18):1951-1961. doi: 10.1182/blood.2024024761.
2
Electroencephalography (EEG) based epilepsy diagnosis via multiple feature space fusion using shared hidden space-driven multi-view learning.基于脑电图(EEG)的癫痫诊断:通过使用共享隐藏空间驱动的多视图学习进行多特征空间融合
PeerJ Comput Sci. 2024 Mar 7;10:e1874. doi: 10.7717/peerj-cs.1874. eCollection 2024.
3
Auto-Weighted Multi-View Discriminative Metric Learning Method With Fisher Discriminative and Global Structure Constraints for Epilepsy EEG Signal Classification.
基于Fisher判别和全局结构约束的自动加权多视图判别度量学习方法用于癫痫脑电信号分类
Front Neurosci. 2020 Sep 29;14:586149. doi: 10.3389/fnins.2020.586149. eCollection 2020.
4
Domain Adaptation and Adaptive Information Fusion for Object Detection on Foggy Days.雾天目标检测中的域自适应和自适应信息融合。
Sensors (Basel). 2018 Sep 30;18(10):3286. doi: 10.3390/s18103286.