• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

直接契合自然:生物和人工神经网络的进化视角。

Direct Fit to Nature: An Evolutionary Perspective on Biological and Artificial Neural Networks.

机构信息

Princeton Neuroscience Institute, Princeton University, Princeton, NJ, USA; Department of Psychology, Princeton University, Princeton, NJ, USA.

Princeton Neuroscience Institute, Princeton University, Princeton, NJ, USA.

出版信息

Neuron. 2020 Feb 5;105(3):416-434. doi: 10.1016/j.neuron.2019.12.002.

DOI:10.1016/j.neuron.2019.12.002
PMID:32027833
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7096172/
Abstract

Evolution is a blind fitting process by which organisms become adapted to their environment. Does the brain use similar brute-force fitting processes to learn how to perceive and act upon the world? Recent advances in artificial neural networks have exposed the power of optimizing millions of synaptic weights over millions of observations to operate robustly in real-world contexts. These models do not learn simple, human-interpretable rules or representations of the world; rather, they use local computations to interpolate over task-relevant manifolds in a high-dimensional parameter space. Counterintuitively, similar to evolutionary processes, over-parameterized models can be simple and parsimonious, as they provide a versatile, robust solution for learning a diverse set of functions. This new family of direct-fit models present a radical challenge to many of the theoretical assumptions in psychology and neuroscience. At the same time, this shift in perspective establishes unexpected links with developmental and ecological psychology.

摘要

进化是一种盲目的适应过程,通过这个过程,生物体适应了它们的环境。大脑是否利用类似的盲目适应过程来学习如何感知和作用于世界?人工神经网络的最新进展揭示了通过优化数百万个突触权重来适应数百万个观察结果的强大能力,从而在现实环境中稳健地运行。这些模型并没有学习简单的、人类可解释的世界规则或表示;相反,它们使用局部计算在高维参数空间中对任务相关流形进行插值。反直觉的是,与进化过程类似,过度参数化的模型可以很简单和简洁,因为它们为学习多样化的函数提供了一种通用的、稳健的解决方案。这组新的直接适应模型对心理学和神经科学中的许多理论假设提出了根本性的挑战。与此同时,这种视角的转变与发展心理学和生态心理学建立了意想不到的联系。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/001d/7096172/5a9b968689d5/nihms-1569075-f0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/001d/7096172/bebeef727a33/nihms-1569075-f0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/001d/7096172/b16feff953be/nihms-1569075-f0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/001d/7096172/85baca639284/nihms-1569075-f0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/001d/7096172/5a9b968689d5/nihms-1569075-f0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/001d/7096172/bebeef727a33/nihms-1569075-f0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/001d/7096172/b16feff953be/nihms-1569075-f0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/001d/7096172/85baca639284/nihms-1569075-f0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/001d/7096172/5a9b968689d5/nihms-1569075-f0004.jpg

相似文献

1
Direct Fit to Nature: An Evolutionary Perspective on Biological and Artificial Neural Networks.直接契合自然:生物和人工神经网络的进化视角。
Neuron. 2020 Feb 5;105(3):416-434. doi: 10.1016/j.neuron.2019.12.002.
2
General differential Hebbian learning: Capturing temporal relations between events in neural networks and the brain.一般微分Hebbian 学习:在神经网络和大脑中捕获事件之间的时间关系。
PLoS Comput Biol. 2018 Aug 28;14(8):e1006227. doi: 10.1371/journal.pcbi.1006227. eCollection 2018 Aug.
3
Theories of Error Back-Propagation in the Brain.大脑中的误差反向传播理论。
Trends Cogn Sci. 2019 Mar;23(3):235-250. doi: 10.1016/j.tics.2018.12.005. Epub 2019 Jan 28.
4
Invertible generalized synchronization: A putative mechanism for implicit learning in neural systems.可逆广义同步:神经系统中内隐学习的一种可能机制。
Chaos. 2020 Jun;30(6):063133. doi: 10.1063/5.0004344.
5
Complex computation from developmental priors.基于发育先验的复杂计算。
Nat Commun. 2023 Apr 19;14(1):2226. doi: 10.1038/s41467-023-37980-1.
6
Behaviorally based modeling and computational approaches to neuroscience.基于行为的神经科学建模与计算方法。
Annu Rev Neurosci. 1993;16:597-623. doi: 10.1146/annurev.ne.16.030193.003121.
7
Local dendritic balance enables learning of efficient representations in networks of spiking neurons.局部树突平衡使网络中的尖峰神经元能够学习有效的表示。
Proc Natl Acad Sci U S A. 2021 Dec 14;118(50). doi: 10.1073/pnas.2021925118.
8
Mirrored STDP Implements Autoencoder Learning in a Network of Spiking Neurons.镜像脉冲时间依赖可塑性在脉冲神经元网络中实现自动编码器学习。
PLoS Comput Biol. 2015 Dec 3;11(12):e1004566. doi: 10.1371/journal.pcbi.1004566. eCollection 2015 Dec.
9
Perturbing low dimensional activity manifolds in spiking neuronal networks.扰乱尖峰神经元网络中的低维活动流形。
PLoS Comput Biol. 2019 May 31;15(5):e1007074. doi: 10.1371/journal.pcbi.1007074. eCollection 2019 May.
10
Representations and generalization in artificial and brain neural networks.人工神经网络和大脑神经网络中的表示与泛化。
Proc Natl Acad Sci U S A. 2024 Jul 2;121(27):e2311805121. doi: 10.1073/pnas.2311805121. Epub 2024 Jun 24.

引用本文的文献

1
Ultrasound-assisted extraction of neuroprotective antioxidants from Ganoderma lucidum.超声辅助从灵芝中提取神经保护抗氧化剂
Ultrason Sonochem. 2025 Aug 25;121:107528. doi: 10.1016/j.ultsonch.2025.107528.
2
Skillful subseasonal soil moisture drought forecasts with deep learning-dynamic models.运用深度学习动态模型进行的精准次季节土壤湿度干旱预测。
Nat Commun. 2025 Aug 12;16(1):7461. doi: 10.1038/s41467-025-62761-3.
3
Dynamic Network Plasticity and Sample Efficiency in Biological Neural Cultures: A Comparative Study with Deep Reinforcement Learning.

本文引用的文献

1
The revolution will not be controlled: natural stimuli in speech neuroscience.这场革命无法被控制:言语神经科学中的自然刺激
Lang Cogn Neurosci. 2018 Jul 22;35(5):573-582. doi: 10.1080/23273798.2018.1499946. eCollection 2020.
2
Linguistic generalization and compositionality in modern artificial neural networks.现代人工神经网络中的语言泛化和组合性。
Philos Trans R Soc Lond B Biol Sci. 2020 Feb 3;375(1791):20190307. doi: 10.1098/rstb.2019.0307. Epub 2019 Dec 16.
3
A deep learning framework for neuroscience.深度学习在神经科学中的应用框架。
生物神经培养中的动态网络可塑性与样本效率:与深度强化学习的比较研究
Cyborg Bionic Syst. 2025 Aug 4;6:0336. doi: 10.34133/cbsystems.0336. eCollection 2025.
4
Text-related functionality and dynamics of visual human pre-frontal activations revealed through neural network convergence.通过神经网络收敛揭示的视觉人类前额叶激活的文本相关功能和动态。
Commun Biol. 2025 Jul 30;8(1):1129. doi: 10.1038/s42003-025-08497-8.
5
Reading comprehension in L1 and L2 readers: neurocomputational mechanisms revealed through large language models.第一语言和第二语言阅读者的阅读理解:通过大语言模型揭示的神经计算机制
NPJ Sci Learn. 2025 Jul 10;10(1):46. doi: 10.1038/s41539-025-00337-y.
6
Wakefulness can be distinguished from general anesthesia and sleep in flies using a massive library of univariate time series analyses.利用大量单变量时间序列分析库,可以区分果蝇的清醒状态与全身麻醉和睡眠状态。
PLoS Biol. 2025 Jul 10;23(7):e3003217. doi: 10.1371/journal.pbio.3003217. eCollection 2025 Jul.
7
Emergence of a temporal processing gradient from naturalistic inputs and network connectivity.从自然主义输入和网络连接中出现时间处理梯度。
Proc Natl Acad Sci U S A. 2025 Jul 15;122(28):e2420105122. doi: 10.1073/pnas.2420105122. Epub 2025 Jul 9.
8
The "Podcast" ECoG dataset for modeling neural activity during natural language comprehension.用于在自然语言理解过程中对神经活动进行建模的“播客”脑电图数据集。
Sci Data. 2025 Jul 3;12(1):1135. doi: 10.1038/s41597-025-05462-2.
9
Shared disbelief and shared belief: Belief and disbelief as drivers of interpersonal neural synchronization during narrative processing.共同的怀疑与共同的信念:叙事加工过程中,信念与怀疑作为人际神经同步的驱动因素
Proc Natl Acad Sci U S A. 2025 Jun 10;122(23):e2422396122. doi: 10.1073/pnas.2422396122. Epub 2025 Jun 5.
10
Comparison of Large Language Model with Aphasia.大语言模型与失语症的比较。
Adv Sci (Weinh). 2025 Jun;12(22):e2414016. doi: 10.1002/advs.202414016. Epub 2025 May 14.
Nat Neurosci. 2019 Nov;22(11):1761-1770. doi: 10.1038/s41593-019-0520-2. Epub 2019 Oct 28.
4
The Life of Behavior.行为的一生。
Neuron. 2019 Oct 9;104(1):25-36. doi: 10.1016/j.neuron.2019.09.017.
5
A critique of pure learning and what artificial neural networks can learn from animal brains.对纯粹学习的批判,以及人工神经网络可以从动物大脑中学到什么。
Nat Commun. 2019 Aug 21;10(1):3770. doi: 10.1038/s41467-019-11786-6.
6
High-dimensional geometry of population responses in visual cortex.群体视觉皮层反应的高维几何结构。
Nature. 2019 Jul;571(7765):361-365. doi: 10.1038/s41586-019-1346-5. Epub 2019 Jun 26.
7
Human-level performance in 3D multiplayer games with population-based reinforcement learning.基于群体强化学习的 3D 多人游戏中的人类水平表现。
Science. 2019 May 31;364(6443):859-865. doi: 10.1126/science.aau6249.
8
Evolving Images for Visual Neurons Using a Deep Generative Network Reveals Coding Principles and Neuronal Preferences.利用深度生成网络为视觉神经元生成演变图像,揭示编码原理和神经元偏好。
Cell. 2019 May 2;177(4):999-1009.e10. doi: 10.1016/j.cell.2019.04.005.
9
Reinforcement Learning, Fast and Slow.强化学习:快与慢。
Trends Cogn Sci. 2019 May;23(5):408-422. doi: 10.1016/j.tics.2019.02.006. Epub 2019 Apr 16.
10
Deep Neural Networks as Scientific Models.深度神经网络作为科学模型。
Trends Cogn Sci. 2019 Apr;23(4):305-317. doi: 10.1016/j.tics.2019.01.009. Epub 2019 Feb 19.