• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

一种深度玻尔兹曼机的有效学习过程。

An efficient learning procedure for deep Boltzmann machines.

机构信息

Department of Statistics, University of Toronto, Toronto, Ontario M5S 3G3, Canada.

出版信息

Neural Comput. 2012 Aug;24(8):1967-2006. doi: 10.1162/NECO_a_00311. Epub 2012 Apr 17.

DOI:10.1162/NECO_a_00311
PMID:22509963
Abstract

We present a new learning algorithm for Boltzmann machines that contain many layers of hidden variables. Data-dependent statistics are estimated using a variational approximation that tends to focus on a single mode, and data-independent statistics are estimated using persistent Markov chains. The use of two quite different techniques for estimating the two types of statistic that enter into the gradient of the log likelihood makes it practical to learn Boltzmann machines with multiple hidden layers and millions of parameters. The learning can be made more efficient by using a layer-by-layer pretraining phase that initializes the weights sensibly. The pretraining also allows the variational inference to be initialized sensibly with a single bottom-up pass. We present results on the MNIST and NORB data sets showing that deep Boltzmann machines learn very good generative models of handwritten digits and 3D objects. We also show that the features discovered by deep Boltzmann machines are a very effective way to initialize the hidden layers of feedforward neural nets, which are then discriminatively fine-tuned.

摘要

我们提出了一种新的学习算法,用于包含多层隐藏变量的玻尔兹曼机。使用变分逼近法估计与数据相关的统计信息,该逼近法倾向于关注单个模式,而使用持久马尔可夫链估计与数据无关的统计信息。使用两种非常不同的技术来估计进入对数似然梯度的两种统计信息类型,使得使用具有多个隐藏层和数百万个参数的玻尔兹曼机进行学习变得切实可行。通过使用逐层预训练阶段,可以更有效地进行学习,该阶段可以合理地初始化权重。预训练还允许通过单个自下而上的传递合理地初始化变分推理。我们在 MNIST 和 NORB 数据集上的结果表明,深度玻尔兹曼机学习手写数字和 3D 物体的非常好的生成模型。我们还表明,深度玻尔兹曼机发现的特征是初始化前馈神经网络隐藏层的一种非常有效的方法,然后可以对其进行有区别的微调。

相似文献

1
An efficient learning procedure for deep Boltzmann machines.一种深度玻尔兹曼机的有效学习过程。
Neural Comput. 2012 Aug;24(8):1967-2006. doi: 10.1162/NECO_a_00311. Epub 2012 Apr 17.
2
Where do features come from?特征从何而来?
Cogn Sci. 2014 Aug;38(6):1078-101. doi: 10.1111/cogs.12049. Epub 2013 Jun 25.
3
A fast learning algorithm for deep belief nets.一种用于深度信念网络的快速学习算法。
Neural Comput. 2006 Jul;18(7):1527-54. doi: 10.1162/neco.2006.18.7.1527.
4
Representational power of restricted boltzmann machines and deep belief networks.受限玻尔兹曼机和深度信念网络的表征能力。
Neural Comput. 2008 Jun;20(6):1631-49. doi: 10.1162/neco.2008.04-07-510.
5
Approximate learning algorithm in Boltzmann machines.玻尔兹曼机中的近似学习算法。
Neural Comput. 2009 Nov;21(11):3130-78. doi: 10.1162/neco.2009.08-08-844.
6
Accelerating deep learning with memcomputing.利用忆阻器计算加速深度学习。
Neural Netw. 2019 Feb;110:1-7. doi: 10.1016/j.neunet.2018.10.012. Epub 2018 Nov 3.
7
Factorial hidden Markov models and the generalized backfitting algorithm.阶乘隐马尔可夫模型与广义反向拟合算法。
Neural Comput. 2002 Oct;14(10):2415-37. doi: 10.1162/08997660260293283.
8
On the equivalence of Hopfield networks and Boltzmann Machines.Hopfield 网络与 Boltzmann 机的等价性。
Neural Netw. 2012 Oct;34:1-9. doi: 10.1016/j.neunet.2012.06.003. Epub 2012 Jun 23.
9
Emergence of Compositional Representations in Restricted Boltzmann Machines.受限玻尔兹曼机中组合表示的出现。
Phys Rev Lett. 2017 Mar 31;118(13):138301. doi: 10.1103/PhysRevLett.118.138301. Epub 2017 Mar 28.
10
Deep, big, simple neural nets for handwritten digit recognition.深度、大型、简单的神经网络用于手写数字识别。
Neural Comput. 2010 Dec;22(12):3207-20. doi: 10.1162/NECO_a_00052. Epub 2010 Sep 21.

引用本文的文献

1
Improving care interactions (and training) in nursing homes with artificial intelligence.利用人工智能改善养老院的护理互动(及培训)。
Geroscience. 2025 Feb 25. doi: 10.1007/s11357-025-01557-1.
2
A Novel Detection Scheme for Motor Bearing Structure Defects in a High-Speed Train Using Stator Current.一种基于定子电流的高速列车电机轴承结构缺陷新型检测方案。
Sensors (Basel). 2024 Nov 30;24(23):7675. doi: 10.3390/s24237675.
3
Deep Learning for Medical Image-Based Cancer Diagnosis.基于医学图像的癌症诊断的深度学习
Cancers (Basel). 2023 Jul 13;15(14):3608. doi: 10.3390/cancers15143608.
4
Estimation of silent phenotypes of calf antibiotic dysbiosis.评估犊牛抗生素相关肠道菌群失调的隐匿表型。
Sci Rep. 2023 Apr 19;13(1):6359. doi: 10.1038/s41598-023-33444-0.
5
Deep learning in economics: a systematic and critical review.经济学中的深度学习:系统与批判性综述
Artif Intell Rev. 2023 Feb 4:1-43. doi: 10.1007/s10462-022-10272-8.
6
Heart sound classification based on improved mel-frequency spectral coefficients and deep residual learning.基于改进的梅尔频率谱系数和深度残差学习的心音分类
Front Physiol. 2022 Dec 22;13:1084420. doi: 10.3389/fphys.2022.1084420. eCollection 2022.
7
Smartphone-based holographic measurement of polydisperse suspended particulate matter with various mass concentration ratios.基于智能手机的具有不同质量浓度比的多分散悬浮颗粒物的全息测量。
Sci Rep. 2022 Dec 30;12(1):22609. doi: 10.1038/s41598-022-27215-6.
8
Breaking away from labels: The promise of self-supervised machine learning in intelligent health.摆脱标签:自监督机器学习在智能健康领域的前景。
Patterns (N Y). 2022 Feb 11;3(2):100410. doi: 10.1016/j.patter.2021.100410.
9
Mode-assisted joint training of deep Boltzmann machines.深度玻尔兹曼机的模式辅助联合训练
Sci Rep. 2021 Sep 24;11(1):19000. doi: 10.1038/s41598-021-98404-y.
10
Integrative omics of schizophrenia: from genetic determinants to clinical classification and risk prediction.精神分裂症的综合组学:从遗传决定因素到临床分类和风险预测。
Mol Psychiatry. 2022 Jan;27(1):113-126. doi: 10.1038/s41380-021-01201-2. Epub 2021 Jun 30.