John A. Paulson School of Engineering and Applied Sciences, Harvard University and Center for Brain Science, Harvard University, United States.
Applied Mathematics, Physiology and Biophysics, and Computational Neuroscience Center, University of Washington, United States.
Curr Opin Neurobiol. 2023 Dec;83:102780. doi: 10.1016/j.conb.2023.102780. Epub 2023 Sep 25.
Neural circuits-both in the brain and in "artificial" neural network models-learn to solve a remarkable variety of tasks, and there is a great current opportunity to use neural networks as models for brain function. Key to this endeavor is the ability to characterize the representations formed by both artificial and biological brains. Here, we investigate this potential through the lens of recently developing theory that characterizes neural networks as "lazy" or "rich" depending on the approach they use to solve tasks: lazy networks solve tasks by making small changes in connectivity, while rich networks solve tasks by significantly modifying weights throughout the network (including "hidden layers"). We further elucidate rich networks through the lens of compression and "neural collapse", ideas that have recently been of significant interest to neuroscience and machine learning. We then show how these ideas apply to a domain of increasing importance to both fields: extracting latent structures through self-supervised learning.
神经回路——无论是在大脑中还是在“人工”神经网络模型中——都学会了解决各种非凡的任务,现在有一个很好的机会可以使用神经网络作为大脑功能的模型。这项工作的关键是能够描述人工和生物大脑所形成的表示。在这里,我们通过最近发展的理论来研究这种可能性,该理论根据网络解决任务的方法将神经网络描述为“懒惰”或“丰富”:懒惰的网络通过在连接性上进行小的改变来解决任务,而丰富的网络则通过在整个网络中显著修改权重(包括“隐藏层”)来解决任务。我们通过压缩和“神经崩溃”的视角进一步阐明了丰富的网络,这些概念最近在神经科学和机器学习领域引起了极大的兴趣。然后,我们展示了这些想法如何应用于这两个领域越来越重要的领域:通过自我监督学习提取潜在结构。