• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

使用大语言模型驱动和抑制人类语言网络。

Driving and suppressing the human language network using large language models.

作者信息

Tuckute Greta, Sathe Aalok, Srikant Shashank, Taliaferro Maya, Wang Mingye, Schrimpf Martin, Kay Kendrick, Fedorenko Evelina

机构信息

Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA 02139 USA.

McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139 USA.

出版信息

bioRxiv. 2023 Oct 30:2023.04.16.537080. doi: 10.1101/2023.04.16.537080.

DOI:10.1101/2023.04.16.537080
PMID:37090673
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10120732/
Abstract

Transformer models such as GPT generate human-like language and are highly predictive of human brain responses to language. Here, using fMRI-measured brain responses to 1,000 diverse sentences, we first show that a GPT-based encoding model can predict the magnitude of brain response associated with each sentence. Then, we use the model to identify new sentences that are predicted to drive or suppress responses in the human language network. We show that these model-selected novel sentences indeed strongly drive and suppress activity of human language areas in new individuals. A systematic analysis of the model-selected sentences reveals that surprisal and well-formedness of linguistic input are key determinants of response strength in the language network. These results establish the ability of neural network models to not only mimic human language but also noninvasively control neural activity in higher-level cortical areas, like the language network.

摘要

像GPT这样的Transformer模型能够生成类人语言,并且对人类大脑对语言的反应具有高度预测性。在此,我们利用功能磁共振成像(fMRI)测量的大脑对1000个不同句子的反应,首先表明基于GPT的编码模型能够预测与每个句子相关的大脑反应强度。然后,我们使用该模型识别预计会驱动或抑制人类语言网络反应的新句子。我们发现,这些模型选择的新句子确实能强烈驱动和抑制新个体中人类语言区域的活动。对模型选择的句子进行系统分析后发现,语言输入的意外性和语法正确性是语言网络中反应强度的关键决定因素。这些结果证明了神经网络模型不仅能够模仿人类语言,还能够非侵入性地控制像语言网络这样的高级皮层区域的神经活动。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/01ea/10621371/f5ac3dad155f/nihpp-2023.04.16.537080v4-f0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/01ea/10621371/75fbfa4f1770/nihpp-2023.04.16.537080v4-f0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/01ea/10621371/3ee4fe88a214/nihpp-2023.04.16.537080v4-f0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/01ea/10621371/620a5b9b6a0c/nihpp-2023.04.16.537080v4-f0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/01ea/10621371/8b2a283c212f/nihpp-2023.04.16.537080v4-f0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/01ea/10621371/4421f9cd45a1/nihpp-2023.04.16.537080v4-f0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/01ea/10621371/f5ac3dad155f/nihpp-2023.04.16.537080v4-f0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/01ea/10621371/75fbfa4f1770/nihpp-2023.04.16.537080v4-f0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/01ea/10621371/3ee4fe88a214/nihpp-2023.04.16.537080v4-f0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/01ea/10621371/620a5b9b6a0c/nihpp-2023.04.16.537080v4-f0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/01ea/10621371/8b2a283c212f/nihpp-2023.04.16.537080v4-f0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/01ea/10621371/4421f9cd45a1/nihpp-2023.04.16.537080v4-f0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/01ea/10621371/f5ac3dad155f/nihpp-2023.04.16.537080v4-f0006.jpg

相似文献

1
Driving and suppressing the human language network using large language models.使用大语言模型驱动和抑制人类语言网络。
bioRxiv. 2023 Oct 30:2023.04.16.537080. doi: 10.1101/2023.04.16.537080.
2
Driving and suppressing the human language network using large language models.使用大型语言模型驱动和抑制人类语言网络。
Nat Hum Behav. 2024 Mar;8(3):544-561. doi: 10.1038/s41562-023-01783-7. Epub 2024 Jan 3.
3
Deep Artificial Neural Networks Reveal a Distributed Cortical Network Encoding Propositional Sentence-Level Meaning.深度人工神经网络揭示命题句级意义的分布式皮层网络编码。
J Neurosci. 2021 May 5;41(18):4100-4119. doi: 10.1523/JNEUROSCI.1152-20.2021. Epub 2021 Mar 22.
4
Neural Encoding and Decoding With Distributed Sentence Representations.分布式句子表示的神经编码和解码。
IEEE Trans Neural Netw Learn Syst. 2021 Feb;32(2):589-603. doi: 10.1109/TNNLS.2020.3027595. Epub 2021 Feb 4.
5
Composition is the Core Driver of the Language-selective Network.成分是语言选择网络的核心驱动因素。
Neurobiol Lang (Camb). 2020 Mar 1;1(1):104-134. doi: 10.1162/nol_a_00005. eCollection 2020.
6
Combined eye tracking and fMRI reveals neural basis of linguistic predictions during sentence comprehension.眼动追踪与功能磁共振成像相结合揭示句子理解过程中语言预测的神经基础。
Cortex. 2015 Jul;68:33-47. doi: 10.1016/j.cortex.2015.04.011. Epub 2015 Apr 27.
7
Delta-Band Neural Responses to Individual Words Are Modulated by Sentence Processing.Delta 波段神经对单个单词的反应受句子处理的调节。
J Neurosci. 2023 Jun 28;43(26):4867-4883. doi: 10.1523/JNEUROSCI.0964-22.2023. Epub 2023 May 23.
8
Lexical semantic content, not syntactic structure, is the main contributor to ANN-brain similarity of fMRI responses in the language network.词汇语义内容而非句法结构,是语言网络中功能磁共振成像反应的人工神经网络与大脑相似性的主要促成因素。
bioRxiv. 2023 May 6:2023.05.05.539646. doi: 10.1101/2023.05.05.539646.
9
Lexical-Semantic Content, Not Syntactic Structure, Is the Main Contributor to ANN-Brain Similarity of fMRI Responses in the Language Network.词汇语义内容而非句法结构是语言网络中功能磁共振成像反应的人工神经网络与大脑相似性的主要贡献因素。
Neurobiol Lang (Camb). 2024 Apr 1;5(1):7-42. doi: 10.1162/nol_a_00116. eCollection 2024.
10
Comparison of Structural Parsers and Neural Language Models as Surprisal Estimators.作为惊奇度估计器的结构解析器与神经语言模型的比较。
Front Artif Intell. 2022 Mar 3;5:777963. doi: 10.3389/frai.2022.777963. eCollection 2022.

本文引用的文献

1
Neural populations in the language network differ in the size of their temporal receptive windows.语言网络中的神经群体在时间感受窗的大小上存在差异。
Nat Hum Behav. 2024 Oct;8(10):1924-1942. doi: 10.1038/s41562-024-01944-2. Epub 2024 Aug 26.
2
The Language Network Reliably "Tracks" Naturalistic Meaningful Nonverbal Stimuli.语言网络能够可靠地“追踪”自然主义的有意义非言语刺激。
Neurobiol Lang (Camb). 2024 Jun 3;5(2):385-408. doi: 10.1162/nol_a_00135. eCollection 2024.
3
Computational Language Modeling and the Promise of In Silico Experimentation.
计算语言建模与计算机模拟实验的前景。
Neurobiol Lang (Camb). 2024 Apr 1;5(1):80-106. doi: 10.1162/nol_a_00101. eCollection 2024.
4
Strong Prediction: Language Model Surprisal Explains Multiple N400 Effects.强预测:语言模型意外值解释多种N400效应。
Neurobiol Lang (Camb). 2024 Apr 1;5(1):107-135. doi: 10.1162/nol_a_00105. eCollection 2024.
5
Artificial Neural Network Language Models Predict Human Brain Responses to Language Even After a Developmentally Realistic Amount of Training.即使经过符合发育实际的训练量,人工神经网络语言模型仍能预测人类大脑对语言的反应。
Neurobiol Lang (Camb). 2024 Apr 1;5(1):43-63. doi: 10.1162/nol_a_00137. eCollection 2024.
6
Predictive Coding or Just Feature Discovery? An Alternative Account of Why Language Models Fit Brain Data.预测编码还是仅仅是特征发现?关于语言模型为何符合大脑数据的另一种解释。
Neurobiol Lang (Camb). 2024 Apr 1;5(1):64-79. doi: 10.1162/nol_a_00087. eCollection 2024.
7
Functional characterization of the language network of polyglots and hyperpolyglots with precision fMRI.利用精确 fMRI 技术对多语者和超语者语言网络的功能特征进行研究。
Cereb Cortex. 2024 Mar 1;34(3). doi: 10.1093/cercor/bhae049.
8
Large-scale evidence for logarithmic effects of word predictability on reading time.大规模证据表明,单词可预测性对阅读时间的影响呈对数关系。
Proc Natl Acad Sci U S A. 2024 Mar 5;121(10):e2307876121. doi: 10.1073/pnas.2307876121. Epub 2024 Feb 29.
9
Event Knowledge in Large Language Models: The Gap Between the Impossible and the Unlikely.大型语言模型中的事件知识:不可能与不太可能之间的差距。
Cogn Sci. 2023 Nov;47(11):e13386. doi: 10.1111/cogs.13386.
10
A social-semantic working-memory account for two canonical language areas.两个经典语言区的社会语义工作记忆解释。
Nat Hum Behav. 2023 Nov;7(11):1980-1997. doi: 10.1038/s41562-023-01704-8. Epub 2023 Sep 21.