• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

特征从何而来?

Where do features come from?

作者信息

Hinton Geoffrey

机构信息

Department of Computer Science, University of Toronto.

出版信息

Cogn Sci. 2014 Aug;38(6):1078-101. doi: 10.1111/cogs.12049. Epub 2013 Jun 25.

DOI:10.1111/cogs.12049
PMID:23800216
Abstract

It is possible to learn multiple layers of non-linear features by backpropagating error derivatives through a feedforward neural network. This is a very effective learning procedure when there is a huge amount of labeled training data, but for many learning tasks very few labeled examples are available. In an effort to overcome the need for labeled data, several different generative models were developed that learned interesting features by modeling the higher order statistical structure of a set of input vectors. One of these generative models, the restricted Boltzmann machine (RBM), has no connections between its hidden units and this makes perceptual inference and learning much simpler. More significantly, after a layer of hidden features has been learned, the activities of these features can be used as training data for another RBM. By applying this idea recursively, it is possible to learn a deep hierarchy of progressively more complicated features without requiring any labeled data. This deep hierarchy can then be treated as a feedforward neural network which can be discriminatively fine-tuned using backpropagation. Using a stack of RBMs to initialize the weights of a feedforward neural network allows backpropagation to work effectively in much deeper networks and it leads to much better generalization. A stack of RBMs can also be used to initialize a deep Boltzmann machine that has many hidden layers. Combining this initialization method with a new method for fine-tuning the weights finally leads to the first efficient way of training Boltzmann machines with many hidden layers and millions of weights.

摘要

通过前馈神经网络反向传播误差导数来学习多层非线性特征是可行的。当有大量带标签的训练数据时,这是一种非常有效的学习方法,但对于许多学习任务来说,可用的带标签示例很少。为了克服对带标签数据的需求,人们开发了几种不同的生成模型,这些模型通过对一组输入向量的高阶统计结构进行建模来学习有趣的特征。其中一种生成模型,即受限玻尔兹曼机(RBM),其隐藏单元之间没有连接,这使得感知推理和学习变得更加简单。更重要的是,在学习了一层隐藏特征之后,这些特征的活动可以用作另一个RBM的训练数据。通过递归地应用这个想法,可以学习到一个越来越复杂的特征的深度层次结构,而无需任何带标签的数据。然后,这个深度层次结构可以被视为一个前馈神经网络,可以使用反向传播进行有区别的微调。使用一堆RBM来初始化前馈神经网络的权重,可以使反向传播在更深的网络中有效地工作,并导致更好的泛化。一堆RBM也可以用来初始化一个有许多隐藏层的深度玻尔兹曼机。将这种初始化方法与一种新的权重微调方法相结合,最终产生了第一种有效训练具有许多隐藏层和数百万权重的玻尔兹曼机的方法。

相似文献

1
Where do features come from?特征从何而来?
Cogn Sci. 2014 Aug;38(6):1078-101. doi: 10.1111/cogs.12049. Epub 2013 Jun 25.
2
An efficient learning procedure for deep Boltzmann machines.一种深度玻尔兹曼机的有效学习过程。
Neural Comput. 2012 Aug;24(8):1967-2006. doi: 10.1162/NECO_a_00311. Epub 2012 Apr 17.
3
Representational power of restricted boltzmann machines and deep belief networks.受限玻尔兹曼机和深度信念网络的表征能力。
Neural Comput. 2008 Jun;20(6):1631-49. doi: 10.1162/neco.2008.04-07-510.
4
On the equivalence of Hopfield networks and Boltzmann Machines.Hopfield 网络与 Boltzmann 机的等价性。
Neural Netw. 2012 Oct;34:1-9. doi: 10.1016/j.neunet.2012.06.003. Epub 2012 Jun 23.
5
Learning Orthographic Structure With Sequential Generative Neural Networks.使用序列生成神经网络学习正字法结构。
Cogn Sci. 2016 Apr;40(3):579-606. doi: 10.1111/cogs.12258. Epub 2015 Jun 14.
6
Expected energy-based restricted Boltzmann machine for classification.预期基于能量的受限玻尔兹曼机分类。
Neural Netw. 2015 Apr;64:29-38. doi: 10.1016/j.neunet.2014.09.006. Epub 2014 Sep 28.
7
An improvement of extreme learning machine for compact single-hidden-layer feedforward neural networks.用于紧凑型单隐层前馈神经网络的极限学习机改进方法。
Int J Neural Syst. 2008 Oct;18(5):433-41. doi: 10.1142/S0129065708001695.
8
Accelerating deep learning with memcomputing.利用忆阻器计算加速深度学习。
Neural Netw. 2019 Feb;110:1-7. doi: 10.1016/j.neunet.2018.10.012. Epub 2018 Nov 3.
9
Dynamical analysis of contrastive divergence learning: Restricted Boltzmann machines with Gaussian visible units.对比散度学习的动力学分析:具有高斯可见单元的受限玻尔兹曼机。
Neural Netw. 2016 Jul;79:78-87. doi: 10.1016/j.neunet.2016.03.013. Epub 2016 Apr 12.
10
A fast learning algorithm for deep belief nets.一种用于深度信念网络的快速学习算法。
Neural Comput. 2006 Jul;18(7):1527-54. doi: 10.1162/neco.2006.18.7.1527.

引用本文的文献

1
Development and evaluation of a java-based deep neural network method for drug response predictions.一种基于Java的用于药物反应预测的深度神经网络方法的开发与评估。
Front Artif Intell. 2023 Mar 23;6:1069353. doi: 10.3389/frai.2023.1069353. eCollection 2023.
2
Comparing biological information contained in mRNA and non-coding RNAs for classification of lung cancer patients.比较 mRNA 和非编码 RNA 中包含的生物学信息,以对肺癌患者进行分类。
BMC Cancer. 2019 Dec 3;19(1):1176. doi: 10.1186/s12885-019-6338-1.
3
Comparing deep belief networks with support vector machines for classifying gene expression data from complex disorders.
比较深度置信网络与支持向量机在分类复杂疾病基因表达数据中的应用。
FEBS Open Bio. 2019 Jul;9(7):1232-1248. doi: 10.1002/2211-5463.12652. Epub 2019 Jun 7.
4
Bayesian cognitive science, predictive brains, and the nativism debate.贝叶斯认知科学、预测性大脑与先天论之争
Synthese. 2018;195(11):4817-4838. doi: 10.1007/s11229-017-1427-7. Epub 2017 May 22.
5
Computational Foundations of Natural Intelligence.自然智能的计算基础
Front Comput Neurosci. 2017 Dec 7;11:112. doi: 10.3389/fncom.2017.00112. eCollection 2017.
6
Analyzing Distributional Learning of Phonemic Categories in Unsupervised Deep Neural Networks.分析无监督深度神经网络中音素类别的分布学习
Cogsci. 2016 Aug;2016:1757-1762.
7
The Role of Architectural and Learning Constraints in Neural Network Models: A Case Study on Visual Space Coding.架构和学习约束在神经网络模型中的作用:视觉空间编码的案例研究
Front Comput Neurosci. 2017 Mar 21;11:13. doi: 10.3389/fncom.2017.00013. eCollection 2017.
8
Spatial generalization in operant learning: lessons from professional basketball.操作性学习中的空间泛化:来自职业篮球的经验教训。
PLoS Comput Biol. 2014 May 22;10(5):e1003623. doi: 10.1371/journal.pcbi.1003623. eCollection 2014 May.
9
Deep generative learning of location-invariant visual word recognition.位置不变视觉词识别的深度生成学习。
Front Psychol. 2013 Sep 19;4:635. doi: 10.3389/fpsyg.2013.00635. eCollection 2013.
10
Modeling language and cognition with deep unsupervised learning: a tutorial overview.基于深度无监督学习的语言和认知建模:教程概述。
Front Psychol. 2013 Aug 20;4:515. doi: 10.3389/fpsyg.2013.00515. eCollection 2013.