Marblestone Adam H, Wayne Greg, Kording Konrad P
Synthetic Neurobiology Group, Massachusetts Institute of Technology, Media Lab Cambridge, MA, USA.
Google Deepmind London, UK.
Front Comput Neurosci. 2016 Sep 14;10:94. doi: 10.3389/fncom.2016.00094. eCollection 2016.
Neuroscience has focused on the detailed implementation of computation, studying neural codes, dynamics and circuits. In machine learning, however, artificial neural networks tend to eschew precisely designed codes, dynamics or circuits in favor of brute force optimization of a cost function, often using simple and relatively uniform initial architectures. Two recent developments have emerged within machine learning that create an opportunity to connect these seemingly divergent perspectives. First, structured architectures are used, including dedicated systems for attention, recursion and various forms of short- and long-term memory storage. Second, cost functions and training procedures have become more complex and are varied across layers and over time. Here we think about the brain in terms of these ideas. We hypothesize that (1) the brain optimizes cost functions, (2) the cost functions are diverse and differ across brain locations and over development, and (3) optimization operates within a pre-structured architecture matched to the computational problems posed by behavior. In support of these hypotheses, we argue that a range of implementations of credit assignment through multiple layers of neurons are compatible with our current knowledge of neural circuitry, and that the brain's specialized systems can be interpreted as enabling efficient optimization for specific problem classes. Such a heterogeneously optimized system, enabled by a series of interacting cost functions, serves to make learning data-efficient and precisely targeted to the needs of the organism. We suggest directions by which neuroscience could seek to refine and test these hypotheses.
神经科学一直专注于计算的详细实现,研究神经编码、动力学和神经回路。然而,在机器学习中,人工神经网络往往避开精心设计的编码、动力学或回路,转而采用对代价函数进行强力优化的方式,通常使用简单且相对统一的初始架构。机器学习领域最近出现了两个新进展,为连接这些看似不同的观点创造了机会。第一,采用了结构化架构,包括用于注意力、递归以及各种形式短期和长期记忆存储的专用系统。第二,代价函数和训练过程变得更加复杂,并且在不同层和不同时间有所变化。在此,我们依据这些观点来思考大脑。我们假设:(1)大脑对代价函数进行优化;(2)代价函数多种多样,在大脑不同位置以及发育过程中存在差异;(3)优化在与行为所带来的计算问题相匹配的预结构化架构内进行。为支持这些假设,我们认为通过多层神经元进行的一系列信用分配实现方式与我们目前对神经回路的认识相符,并且大脑的专门系统可以被解释为能够针对特定问题类别实现高效优化。这样一个由一系列相互作用的代价函数促成的异质优化系统,有助于使学习在数据利用上更加高效,并且能够精确地针对生物体的需求。我们提出了神经科学可以用来完善和检验这些假设的方向。