• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

大规模字典学习的随机算法。

Randomized algorithms for large-scale dictionary learning.

机构信息

School of Mathematics, China University of Mining and Technology, Xuzhou, 221116, Jiangsu, PR China; School of Big Data, Fuzhou University of International Studies and Trade, Fuzhou, Fujian, PR China.

School of Mathematics, China University of Mining and Technology, Xuzhou, 221116, Jiangsu, PR China.

出版信息

Neural Netw. 2024 Nov;179:106628. doi: 10.1016/j.neunet.2024.106628. Epub 2024 Aug 10.

DOI:10.1016/j.neunet.2024.106628
PMID:39168071
Abstract

Dictionary learning is an important sparse representation algorithm which has been widely used in machine learning and artificial intelligence. However, for massive data in the big data era, classical dictionary learning algorithms are computationally expensive and even can be infeasible. To overcome this difficulty, we propose new dictionary learning methods based on randomized algorithms. The contributions of this work are as follows. First, we find that dictionary matrix is often numerically low-rank. Based on this property, we apply randomized singular value decomposition (RSVD) to the dictionary matrix, and propose a randomized algorithm for linear dictionary learning. Compared with the classical K-SVD algorithm, an advantage is that one can update all the elements of the dictionary matrix simultaneously. Second, to the best of our knowledge, there are few theoretical results on why one can solve the involved matrix computation problems inexactly in dictionary learning. To fill-in this gap, we show the rationality of this randomized algorithm with inexact solving, from a matrix perturbation analysis point of view. Third, based on the numerically low-rank property and Nyström approximation of the kernel matrix, we propose a randomized kernel dictionary learning algorithm, and establish the distance between the exact solution and the computed solution, to show the effectiveness of the proposed randomized kernel dictionary learning algorithm. Fourth, we propose an efficient scheme for the testing stage in kernel dictionary learning. By using this strategy, there is no need to form nor store kernel matrices explicitly both in the training and the testing stages. Comprehensive numerical experiments are performed on some real-world data sets. Numerical results demonstrate the rationality of our strategies, and show that the proposed algorithms are much efficient than some state-of-the-art dictionary learning algorithms. The MATLAB codes of the proposed algorithms are publicly available from https://github.com/Jiali-yang/RALDL_RAKDL.

摘要

字典学习是一种重要的稀疏表示算法,已被广泛应用于机器学习和人工智能领域。然而,对于大数据时代的海量数据,经典的字典学习算法计算量很大,甚至可能不可行。为了克服这一困难,我们提出了基于随机算法的新字典学习方法。这项工作的贡献如下。首先,我们发现字典矩阵通常在数值上是低秩的。基于这一特性,我们将随机奇异值分解(RSVD)应用于字典矩阵,并提出了一种用于线性字典学习的随机算法。与经典的 K-SVD 算法相比,一个优点是可以同时更新字典矩阵的所有元素。其次,据我们所知,关于为什么在字典学习中可以不精确地解决所涉及的矩阵计算问题,几乎没有理论结果。为了填补这一空白,我们从矩阵扰动分析的角度展示了这种随机算法的合理性。第三,基于核矩阵的数值低秩特性和 Nyström 逼近,我们提出了一种随机核字典学习算法,并建立了精确解和计算解之间的距离,以显示所提出的随机核字典学习算法的有效性。第四,我们提出了一种核字典学习中测试阶段的有效方案。通过使用这种策略,在训练和测试阶段都不需要显式地形成和存储核矩阵。在一些真实数据集上进行了全面的数值实验。数值结果证明了我们策略的合理性,并表明所提出的算法比一些最先进的字典学习算法效率更高。所提出算法的 MATLAB 代码可在 https://github.com/Jiali-yang/RALDL_RAKDL 上公开获取。

相似文献

1
Randomized algorithms for large-scale dictionary learning.大规模字典学习的随机算法。
Neural Netw. 2024 Nov;179:106628. doi: 10.1016/j.neunet.2024.106628. Epub 2024 Aug 10.
2
Large-scale Nyström kernel matrix approximation using randomized SVD.基于随机奇异值分解的大规模 Nyström 核矩阵逼近。
IEEE Trans Neural Netw Learn Syst. 2015 Jan;26(1):152-64. doi: 10.1109/TNNLS.2014.2359798. Epub 2014 Oct 8.
3
An analysis dictionary learning algorithm under a noisy data model with orthogonality constraint.一种具有正交性约束的噪声数据模型下的分析字典学习算法。
ScientificWorldJournal. 2014;2014:852978. doi: 10.1155/2014/852978. Epub 2014 Jul 13.
4
Clustered Nyström method for large scale manifold learning and dimension reduction.用于大规模流形学习和降维的聚类奈斯特罗姆方法。
IEEE Trans Neural Netw. 2010 Oct;21(10):1576-87. doi: 10.1109/TNN.2010.2064786. Epub 2010 Aug 30.
5
Orthogonal Procrustes Analysis for Dictionary Learning in Sparse Linear Representation.稀疏线性表示中用于字典学习的正交普罗克汝斯分析
PLoS One. 2017 Jan 19;12(1):e0169663. doi: 10.1371/journal.pone.0169663. eCollection 2017.
6
Sequential Dictionary Learning From Correlated Data: Application to fMRI Data Analysis.基于相关数据的序贯字典学习:在 fMRI 数据分析中的应用。
IEEE Trans Image Process. 2017 Jun;26(6):3002-3015. doi: 10.1109/TIP.2017.2686014. Epub 2017 Mar 22.
7
Three layered sparse dictionary learning algorithm for enhancing the subject wise segregation of brain networks.用于增强脑网络主题间分离的三层稀疏字典学习算法。
Sci Rep. 2024 Aug 17;14(1):19070. doi: 10.1038/s41598-024-69647-2.
8
An efficient dictionary learning algorithm and its application to 3-D medical image denoising.一种高效的字典学习算法及其在三维医学图像去噪中的应用。
IEEE Trans Biomed Eng. 2012 Feb;59(2):417-27. doi: 10.1109/TBME.2011.2173935. Epub 2011 Oct 27.
9
Learning Low-Rank Class-Specific Dictionary and Sparse Intra-Class Variant Dictionary for Face Recognition.学习用于人脸识别的低秩特定类别字典和稀疏类内变体字典。
PLoS One. 2015 Nov 16;10(11):e0142403. doi: 10.1371/journal.pone.0142403. eCollection 2015.
10
Medical image fusion by sparse-based modified fusion framework using block total least-square update dictionary learning algorithm.基于稀疏的改进融合框架的医学图像融合,采用块总体最小二乘更新字典学习算法。
J Med Imaging (Bellingham). 2022 Sep;9(5):052403. doi: 10.1117/1.JMI.9.5.052403. Epub 2022 May 26.