• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

大脑和算法在自然语言处理中部分融合。

Brains and algorithms partially converge in natural language processing.

机构信息

Facebook AI Research, Paris, France.

Université Paris-Saclay, Inria, CEA, Palaiseau, France.

出版信息

Commun Biol. 2022 Feb 16;5(1):134. doi: 10.1038/s42003-022-03036-1.

DOI:10.1038/s42003-022-03036-1
PMID:35173264
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8850612/
Abstract

Deep learning algorithms trained to predict masked words from large amount of text have recently been shown to generate activations similar to those of the human brain. However, what drives this similarity remains currently unknown. Here, we systematically compare a variety of deep language models to identify the computational principles that lead them to generate brain-like representations of sentences. Specifically, we analyze the brain responses to 400 isolated sentences in a large cohort of 102 subjects, each recorded for two hours with functional magnetic resonance imaging (fMRI) and magnetoencephalography (MEG). We then test where and when each of these algorithms maps onto the brain responses. Finally, we estimate how the architecture, training, and performance of these models independently account for the generation of brain-like representations. Our analyses reveal two main findings. First, the similarity between the algorithms and the brain primarily depends on their ability to predict words from context. Second, this similarity reveals the rise and maintenance of perceptual, lexical, and compositional representations within each cortical region. Overall, this study shows that modern language algorithms partially converge towards brain-like solutions, and thus delineates a promising path to unravel the foundations of natural language processing.

摘要

最近的研究表明,经过大量文本训练的深度学习算法可以生成与人类大脑相似的激活模式。然而,目前尚不清楚是什么驱动了这种相似性。在这里,我们系统地比较了各种深度语言模型,以确定导致它们生成类似大脑的句子表示的计算原理。具体来说,我们分析了 102 名受试者中 400 个孤立句子的大脑反应,每个句子都使用功能磁共振成像 (fMRI) 和脑磁图 (MEG) 记录了两个小时。然后,我们测试这些算法中的每一个在何时何地映射到大脑反应上。最后,我们估计这些模型的架构、训练和性能如何独立解释大脑样表示的生成。我们的分析揭示了两个主要发现。首先,算法与大脑之间的相似性主要取决于它们从上下文预测单词的能力。其次,这种相似性揭示了每个皮质区域内感知、词汇和组合表示的出现和维持。总的来说,这项研究表明,现代语言算法部分趋向于类似大脑的解决方案,因此为揭示自然语言处理的基础提供了一条有前途的途径。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4951/8850612/2100659d7743/42003_2022_3036_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4951/8850612/f1ceb8c9dcea/42003_2022_3036_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4951/8850612/1e02905bd892/42003_2022_3036_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4951/8850612/460606b9ed31/42003_2022_3036_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4951/8850612/2100659d7743/42003_2022_3036_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4951/8850612/f1ceb8c9dcea/42003_2022_3036_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4951/8850612/1e02905bd892/42003_2022_3036_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4951/8850612/460606b9ed31/42003_2022_3036_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4951/8850612/2100659d7743/42003_2022_3036_Fig4_HTML.jpg

相似文献

1
Brains and algorithms partially converge in natural language processing.大脑和算法在自然语言处理中部分融合。
Commun Biol. 2022 Feb 16;5(1):134. doi: 10.1038/s42003-022-03036-1.
2
Neural Encoding and Decoding With Distributed Sentence Representations.分布式句子表示的神经编码和解码。
IEEE Trans Neural Netw Learn Syst. 2021 Feb;32(2):589-603. doi: 10.1109/TNNLS.2020.3027595. Epub 2021 Feb 4.
3
Evidence of a predictive coding hierarchy in the human brain listening to speech.人类大脑在听语音时存在预测编码层级的证据。
Nat Hum Behav. 2023 Mar;7(3):430-441. doi: 10.1038/s41562-022-01516-2. Epub 2023 Mar 2.
4
Deep Artificial Neural Networks Reveal a Distributed Cortical Network Encoding Propositional Sentence-Level Meaning.深度人工神经网络揭示命题句级意义的分布式皮层网络编码。
J Neurosci. 2021 May 5;41(18):4100-4119. doi: 10.1523/JNEUROSCI.1152-20.2021. Epub 2021 Mar 22.
5
Individual word representations dissociate from linguistic context along a cortical unimodal to heteromodal gradient.个体单词的表示形式沿着皮质单模态到异模态梯度与语言语境分离。
Hum Brain Mapp. 2024 Feb 1;45(2):e26607. doi: 10.1002/hbm.26607.
6
Lexical-Semantic Content, Not Syntactic Structure, Is the Main Contributor to ANN-Brain Similarity of fMRI Responses in the Language Network.词汇语义内容而非句法结构是语言网络中功能磁共振成像反应的人工神经网络与大脑相似性的主要贡献因素。
Neurobiol Lang (Camb). 2024 Apr 1;5(1):7-42. doi: 10.1162/nol_a_00116. eCollection 2024.
7
Deep language algorithms predict semantic comprehension from brain activity.深度语言算法可以根据大脑活动预测语义理解。
Sci Rep. 2022 Sep 29;12(1):16327. doi: 10.1038/s41598-022-20460-9.
8
Neural correlates of word representation vectors in natural language processing models: Evidence from representational similarity analysis of event-related brain potentials.自然语言处理模型中单词表示向量的神经关联:事件相关脑电位的代表性相似性分析证据。
Psychophysiology. 2022 Mar;59(3):e13976. doi: 10.1111/psyp.13976. Epub 2021 Nov 24.
9
Shared functional specialization in transformer-based language models and the human brain.基于变压器的语言模型和人类大脑的功能专业化共享。
Nat Commun. 2024 Jun 29;15(1):5523. doi: 10.1038/s41467-024-49173-5.
10
Neural representations for newly learned words are modulated by overnight consolidation, reading skill, and age.新学单词的神经表征受夜间巩固、阅读技能和年龄的调节。
Neuropsychologia. 2018 Mar;111:133-144. doi: 10.1016/j.neuropsychologia.2018.01.011. Epub 2018 Jan 31.

引用本文的文献

1
Evidence for compositionality in fMRI visual representations via Brain Algebra.通过脑代数证明功能磁共振成像视觉表征中的组合性。
Commun Biol. 2025 Aug 22;8(1):1263. doi: 10.1038/s42003-025-08706-4.
2
A systematic evaluation of Dutch large language models' surprisal estimates in sentence, paragraph and book reading.对荷兰大语言模型在句子、段落和书籍阅读中的意外度估计进行的系统评估。
Behav Res Methods. 2025 Aug 18;57(9):266. doi: 10.3758/s13428-025-02774-4.
3
The Voxelwise Encoding Model framework: A tutorial introduction to fitting encoding models to fMRI data.

本文引用的文献

1
Spoken language processing activates the primary visual cortex.口语处理激活初级视觉皮层。
PLoS One. 2023 Aug 11;18(8):e0289671. doi: 10.1371/journal.pone.0289671. eCollection 2023.
2
Traces of Meaning Itself: Encoding Distributional Word Vectors in Brain Activity.意义本身的痕迹:在大脑活动中编码分布式词向量
Neurobiol Lang (Camb). 2020 Mar 1;1(1):54-76. doi: 10.1162/nol_a_00003. eCollection 2020.
3
Predictive coding across the left fronto-temporal hierarchy during language comprehension.语言理解过程中左额颞叶层次结构的预测编码。
体素编码模型框架:将编码模型拟合到功能磁共振成像数据的教程介绍。
Imaging Neurosci (Camb). 2025 May 9;3. doi: 10.1162/imag_a_00575. eCollection 2025.
4
Semantic composition in experimental and naturalistic paradigms.实验范式和自然主义范式中的语义合成
Imaging Neurosci (Camb). 2024 Jan 22;2. doi: 10.1162/imag_a_00072. eCollection 2024.
5
Through their eyes: Multi-subject brain decoding with simple alignment techniques.透过他们的眼睛:使用简单对齐技术的多主体脑解码
Imaging Neurosci (Camb). 2024 May 8;2. doi: 10.1162/imag_a_00170. eCollection 2024.
6
Ultralow energy adaptive neuromorphic computing using reconfigurable zinc phosphorus trisulfide memristors.使用可重构三硫化锌磷忆阻器的超低能量自适应神经形态计算
Nat Commun. 2025 Jul 26;16(1):6899. doi: 10.1038/s41467-025-62306-8.
7
Reading comprehension in L1 and L2 readers: neurocomputational mechanisms revealed through large language models.第一语言和第二语言阅读者的阅读理解:通过大语言模型揭示的神经计算机制
NPJ Sci Learn. 2025 Jul 10;10(1):46. doi: 10.1038/s41539-025-00337-y.
8
Cortical language areas are coupled via a soft hierarchy of model-based linguistic features.皮质语言区域通过基于模型的语言特征的软层次结构相互耦合。
bioRxiv. 2025 Jun 3:2025.06.02.657491. doi: 10.1101/2025.06.02.657491.
9
Low-Rank Tensor Encoding Models Decompose Natural Speech Comprehension Processes.低秩张量编码模型分解自然语音理解过程。
bioRxiv. 2025 Jun 3:2025.06.02.657514. doi: 10.1101/2025.06.02.657514.
10
Divergences between Language Models and Human Brains.语言模型与人类大脑之间的差异。
Adv Neural Inf Process Syst. 2024;37:137999-138031.
Cereb Cortex. 2023 Apr 4;33(8):4478-4497. doi: 10.1093/cercor/bhac356.
4
A hierarchy of linguistic predictions during natural language comprehension.自然语言理解过程中的语言预测层次。
Proc Natl Acad Sci U S A. 2022 Aug 9;119(32):e2201968119. doi: 10.1073/pnas.2201968119. Epub 2022 Aug 3.
5
The neural architecture of language: Integrative modeling converges on predictive processing.语言的神经结构:综合建模趋向于预测处理。
Proc Natl Acad Sci U S A. 2021 Nov 9;118(45). doi: 10.1073/pnas.2105646118.
6
The "Narratives" fMRI dataset for evaluating models of naturalistic language comprehension.用于评估自然语言理解模型的“叙事” fMRI 数据集。
Sci Data. 2021 Sep 28;8(1):250. doi: 10.1038/s41597-021-01033-3.
7
Word meaning in minds and machines.思维与机器中的词义。
Psychol Rev. 2023 Mar;130(2):401-431. doi: 10.1037/rev0000297. Epub 2021 Jul 22.
8
Anticipation of temporally structured events in the brain.大脑对时间结构事件的预期。
Elife. 2021 Apr 22;10:e64972. doi: 10.7554/eLife.64972.
9
Spatiotemporal dynamics of orthographic and lexical processing in the ventral visual pathway.腹侧视觉通路上的字形和词汇处理的时空动态。
Nat Hum Behav. 2021 Mar;5(3):389-398. doi: 10.1038/s41562-020-00982-w. Epub 2020 Nov 30.
10
If deep learning is the answer, what is the question?如果深度学习是答案,那么问题是什么?
Nat Rev Neurosci. 2021 Jan;22(1):55-67. doi: 10.1038/s41583-020-00395-8. Epub 2020 Nov 16.