• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

将科学理论评估为语言神经科学中的预测模型。

Evaluating scientific theories as predictive models in language neuroscience.

作者信息

Singh Chandan, Antonello Richard J, Guo Sihang, Mischler Gavin, Gao Jianfeng, Mesgarani Nima, Huth Alexander G

机构信息

Microsoft Research, Redmond, WA, USA.

Electrical Engineering Department, Columbia University, NY, USA.

出版信息

bioRxiv. 2025 Aug 12:2025.08.12.669958. doi: 10.1101/2025.08.12.669958.

DOI:10.1101/2025.08.12.669958
PMID:40832215
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12363820/
Abstract

Modern data-driven encoding models are highly effective at predicting brain responses to language stimuli. However, these models struggle to the underlying phenomena, i.e. what features of the stimulus drive the response? We present Question Answering encoding models, a method for converting qualitative theories of language selectivity into highly accurate, interpretable models of brain responses. QA encoding models annotate a language stimulus by using a large language model to answer yes-no questions corresponding to qualitative theories. A compact QA encoding model that uses only 35 questions outperforms existing baselines at predicting brain responses in both fMRI and ECoG data. The model weights also provide easily interpretable maps of language selectivity across cortex; these maps show quantitative agreement with meta-analyses of the existing literature and selectivity maps identified in a follow-up fMRI experiment. These results demonstrate that LLMs can bridge the widening gap between qualitative scientific theories and data-driven models.

摘要

现代数据驱动的编码模型在预测大脑对语言刺激的反应方面非常有效。然而,这些模型难以理解潜在现象,即刺激的哪些特征驱动了反应?我们提出了问答编码模型,这是一种将语言选择性的定性理论转化为高度准确、可解释的大脑反应模型的方法。问答编码模型通过使用大语言模型回答与定性理论相对应的是非问题来注释语言刺激。一个仅使用35个问题的紧凑问答编码模型在预测功能磁共振成像(fMRI)和皮层脑电图(ECoG)数据中的大脑反应方面优于现有的基线模型。模型权重还提供了整个皮层易于解释的语言选择性图谱;这些图谱显示与现有文献的荟萃分析以及后续fMRI实验中确定的选择性图谱在数量上一致。这些结果表明,大语言模型可以弥合定性科学理论与数据驱动模型之间不断扩大的差距。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2045/12363820/e94b7f29c47e/nihpp-2025.08.12.669958v1-f0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2045/12363820/f42d37732957/nihpp-2025.08.12.669958v1-f0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2045/12363820/779807f86b43/nihpp-2025.08.12.669958v1-f0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2045/12363820/4f51186d2180/nihpp-2025.08.12.669958v1-f0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2045/12363820/e94b7f29c47e/nihpp-2025.08.12.669958v1-f0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2045/12363820/f42d37732957/nihpp-2025.08.12.669958v1-f0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2045/12363820/779807f86b43/nihpp-2025.08.12.669958v1-f0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2045/12363820/4f51186d2180/nihpp-2025.08.12.669958v1-f0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/2045/12363820/e94b7f29c47e/nihpp-2025.08.12.669958v1-f0004.jpg

相似文献

1
Evaluating scientific theories as predictive models in language neuroscience.将科学理论评估为语言神经科学中的预测模型。
bioRxiv. 2025 Aug 12:2025.08.12.669958. doi: 10.1101/2025.08.12.669958.
2
Prescription of Controlled Substances: Benefits and Risks管制药品的处方:益处与风险
3
Short-Term Memory Impairment短期记忆障碍
4
Maternal and neonatal outcomes of elective induction of labor.择期引产的母婴结局
Evid Rep Technol Assess (Full Rep). 2009 Mar(176):1-257.
5
Stigma Management Strategies of Autistic Social Media Users.自闭症社交媒体用户的污名管理策略
Autism Adulthood. 2025 May 28;7(3):273-282. doi: 10.1089/aut.2023.0095. eCollection 2025 Jun.
6
Comparison of Two Modern Survival Prediction Tools, SORG-MLA and METSSS, in Patients With Symptomatic Long-bone Metastases Who Underwent Local Treatment With Surgery Followed by Radiotherapy and With Radiotherapy Alone.两种现代生存预测工具 SORG-MLA 和 METSSS 在接受手术联合放疗和单纯放疗治疗有症状长骨转移患者中的比较。
Clin Orthop Relat Res. 2024 Dec 1;482(12):2193-2208. doi: 10.1097/CORR.0000000000003185. Epub 2024 Jul 23.
7
"I Don't Understand Their Sense of Belonging": Exploring How Nonbinary Autistic Adults Experience Gender.“我不理解他们的归属感”:探索非二元性别的自闭症成年人如何体验性别。
Autism Adulthood. 2024 Dec 2;6(4):462-473. doi: 10.1089/aut.2023.0071. eCollection 2024 Dec.
8
The quantity, quality and findings of network meta-analyses evaluating the effectiveness of GLP-1 RAs for weight loss: a scoping review.评估胰高血糖素样肽-1受体激动剂(GLP-1 RAs)减肥效果的网状Meta分析的数量、质量及结果:一项范围综述
Health Technol Assess. 2025 Jun 25:1-73. doi: 10.3310/SKHT8119.
9
How to Implement Digital Clinical Consultations in UK Maternity Care: the ARM@DA Realist Review.如何在英国产科护理中实施数字临床会诊:ARM@DA实证主义综述
Health Soc Care Deliv Res. 2025 May 21:1-77. doi: 10.3310/WQFV7425.
10
Effectiveness and cost-effectiveness of computer and other electronic aids for smoking cessation: a systematic review and network meta-analysis.计算机和其他电子戒烟辅助手段的有效性和成本效益:系统评价和网络荟萃分析。
Health Technol Assess. 2012;16(38):1-205, iii-v. doi: 10.3310/hta16380.

本文引用的文献

1
The "Podcast" ECoG dataset for modeling neural activity during natural language comprehension.用于在自然语言理解过程中对神经活动进行建模的“播客”脑电图数据集。
Sci Data. 2025 Jul 3;12(1):1135. doi: 10.1038/s41597-025-05462-2.
2
Crafting Interpretable Embeddings for Language Neuroscience by Asking LLMs Questions.通过向语言模型提问来构建用于语言神经科学的可解释嵌入。
Adv Neural Inf Process Syst. 2024;37:124137-124162.
3
Scaling laws for language encoding models in fMRI.功能磁共振成像中语言编码模型的标度律
Adv Neural Inf Process Syst. 2023;36:21895-21907.
4
Predictive Coding or Just Feature Discovery? An Alternative Account of Why Language Models Fit Brain Data.预测编码还是仅仅是特征发现?关于语言模型为何符合大脑数据的另一种解释。
Neurobiol Lang (Camb). 2024 Apr 1;5(1):64-79. doi: 10.1162/nol_a_00087. eCollection 2024.
5
The cortical representation of language timescales is shared between reading and listening.语言的皮质代表时间尺度在阅读和听力之间是共享的。
Commun Biol. 2024 Mar 7;7(1):284. doi: 10.1038/s42003-024-05909-z.
6
Augmenting interpretable models with large language models during training.在训练期间使用大型语言模型增强可解释模型。
Nat Commun. 2023 Nov 30;14(1):7913. doi: 10.1038/s41467-023-43713-1.
7
A natural language fMRI dataset for voxelwise encoding models.基于体素的编码模型的自然语言 fMRI 数据集。
Sci Data. 2023 Aug 23;10(1):555. doi: 10.1038/s41597-023-02437-z.
8
Semantic reconstruction of continuous language from non-invasive brain recordings.从非侵入性脑记录中重建连续语言的语义。
Nat Neurosci. 2023 May;26(5):858-866. doi: 10.1038/s41593-023-01304-9. Epub 2023 May 1.
9
Evidence of a predictive coding hierarchy in the human brain listening to speech.人类大脑在听语音时存在预测编码层级的证据。
Nat Hum Behav. 2023 Mar;7(3):430-441. doi: 10.1038/s41562-022-01516-2. Epub 2023 Mar 2.
10
Shared computational principles for language processing in humans and deep language models.人类和深度语言模型语言处理的共享计算原则。
Nat Neurosci. 2022 Mar;25(3):369-380. doi: 10.1038/s41593-022-01026-4. Epub 2022 Mar 7.