Computational Cognitive Neuroscience Lab, Department of General Psychology, University of Padova Padova, Italy ; IRCCS San Camillo Neurorehabilitation Hospital Venice-Lido, Italy.
Front Psychol. 2013 Aug 20;4:515. doi: 10.3389/fpsyg.2013.00515. eCollection 2013.
Deep unsupervised learning in stochastic recurrent neural networks with many layers of hidden units is a recent breakthrough in neural computation research. These networks build a hierarchy of progressively more complex distributed representations of the sensory data by fitting a hierarchical generative model. In this article we discuss the theoretical foundations of this approach and we review key issues related to training, testing and analysis of deep networks for modeling language and cognitive processing. The classic letter and word perception problem of McClelland and Rumelhart (1981) is used as a tutorial example to illustrate how structured and abstract representations may emerge from deep generative learning. We argue that the focus on deep architectures and generative (rather than discriminative) learning represents a crucial step forward for the connectionist modeling enterprise, because it offers a more plausible model of cortical learning as well as a way to bridge the gap between emergentist connectionist models and structured Bayesian models of cognition.
深度无监督学习在具有多层隐藏单元的随机递归神经网络中是神经计算研究的一个最新突破。这些网络通过拟合分层生成模型,构建了一个逐步更复杂的感觉数据分布式表示的层次结构。在本文中,我们讨论了这种方法的理论基础,并回顾了与语言和认知处理建模的深层网络的训练、测试和分析相关的关键问题。麦克莱兰和鲁梅尔哈特(1981)的经典字母和单词感知问题被用作说明性示例,以说明如何从深度生成学习中产生结构化和抽象的表示。我们认为,关注深度架构和生成(而不是判别)学习是连接主义建模企业向前迈出的关键一步,因为它提供了一种更合理的皮质学习模型,以及一种在突现连接主义模型和认知的结构化贝叶斯模型之间架起桥梁的方法。