• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

人类和大语言模型中潜在树状结构句子表征的积极应用。

Active use of latent tree-structured sentence representation in humans and large language models.

作者信息

Liu Wei, Xiang Ming, Ding Nai

机构信息

Key Laboratory for Biomedical Engineering of Ministry of Education, College of Biomedical Engineering and Instrument Sciences, Zhejiang University, Hangzhou, China.

Department of Linguistics, The University of Chicago, Chicago, IL, USA.

出版信息

Nat Hum Behav. 2025 Sep 10. doi: 10.1038/s41562-025-02297-0.

DOI:10.1038/s41562-025-02297-0
PMID:40931087
Abstract

Understanding how sentences are represented in the human brain, as well as in large language models (LLMs), poses a substantial challenge for cognitive science. Here we develop a one-shot learning task to investigate whether humans and LLMs encode tree-structured constituents within sentences. Participants (total N = 372, native Chinese or English speakers, and bilingual in Chinese and English) and LLMs (for example, ChatGPT) were asked to infer which words should be deleted from a sentence. Both groups tend to delete constituents, instead of non-constituent word strings, following rules specific to Chinese and English, respectively. The results cannot be explained by models that rely only on word properties and word positions. Crucially, based on word strings deleted by either humans or LLMs, the underlying constituency tree structure can be successfully reconstructed. Altogether, these results demonstrate that latent tree-structured sentence representations emerge in both humans and LLMs.

摘要

理解句子在人类大脑以及大语言模型(LLMs)中是如何被表征的,对认知科学来说是一项重大挑战。在此,我们开发了一项一次性学习任务,以研究人类和大语言模型是否在句子中编码了树形结构成分。参与者(总共N = 372人,以中文或英文为母语,且具备中英双语能力)和大语言模型(例如ChatGPT)被要求推断句子中应该删除哪些单词。两组都倾向于分别按照中文和英文特有的规则删除成分,而不是非成分单词串。这些结果无法用仅依赖单词属性和单词位置的模型来解释。至关重要的是,基于人类或大语言模型删除的单词串,可以成功重建潜在的成分树结构。总之,这些结果表明,人类和大语言模型中都出现了潜在的树形结构句子表征。

相似文献

1
Active use of latent tree-structured sentence representation in humans and large language models.人类和大语言模型中潜在树状结构句子表征的积极应用。
Nat Hum Behav. 2025 Sep 10. doi: 10.1038/s41562-025-02297-0.
2
Prescription of Controlled Substances: Benefits and Risks管制药品的处方:益处与风险
3
Large Language Models and Empathy: Systematic Review.大语言模型与同理心:系统综述
J Med Internet Res. 2024 Dec 11;26:e52597. doi: 10.2196/52597.
4
Stench of Errors or the Shine of Potential: The Challenge of (Ir)Responsible Use of ChatGPT in Speech-Language Pathology.错误的恶臭还是潜力的光辉:言语病理学中(不)负责任地使用ChatGPT的挑战。
Int J Lang Commun Disord. 2025 Jul-Aug;60(4):e70088. doi: 10.1111/1460-6984.70088.
5
Aspects of Genetic Diversity, Host Specificity and Public Health Significance of Single-Celled Intestinal Parasites Commonly Observed in Humans and Mostly Referred to as 'Non-Pathogenic'.人类常见且大多被称为“非致病性”的单细胞肠道寄生虫的遗传多样性、宿主特异性及公共卫生意义
APMIS. 2025 Sep;133(9):e70036. doi: 10.1111/apm.70036.
6
Plug-and-play use of tree-based methods: consequences for clinical prediction modeling.基于树的方法的即插即用:对临床预测模型的影响。
J Clin Epidemiol. 2025 Aug;184:111834. doi: 10.1016/j.jclinepi.2025.111834. Epub 2025 May 19.
7
The development of a novel, standardized, norm-referenced Arabic Discourse Assessment Tool (ADAT), including an examination of psychometric properties of discourse measures in aphasia.开发一种新型、标准化、基于常模的阿拉伯语语篇评估工具(ADAT),包括评估失语症患者语篇测量的心理测量特性。
Int J Lang Commun Disord. 2024 Sep-Oct;59(5):2103-2117. doi: 10.1111/1460-6984.13083. Epub 2024 Jun 18.
8
The agreement of phonetic transcriptions between paediatric speech and language therapists transcribing a disordered speech sample.儿科言语和语言治疗师转写语音样本的音标转录的一致性。
Int J Lang Commun Disord. 2024 Sep-Oct;59(5):1981-1995. doi: 10.1111/1460-6984.13043. Epub 2024 Jun 8.
9
Short-Term Memory Impairment短期记忆障碍
10
Unlocking the Secrets Behind Advanced Artificial Intelligence Language Models in Deidentifying Chinese-English Mixed Clinical Text: Development and Validation Study.揭开高级人工智能语言模型在去识别汉英混合临床文本背后的秘密:开发与验证研究。
J Med Internet Res. 2024 Jan 25;26:e48443. doi: 10.2196/48443.

本文引用的文献

1
Predicting the next sentence (not word) in large language models: What model-brain alignment tells us about discourse comprehension.预测大型语言模型中的下一句话(不是下一个词):模型-大脑对齐告诉我们关于语篇理解的什么。
Sci Adv. 2024 May 24;10(21):eadn7744. doi: 10.1126/sciadv.adn7744. Epub 2024 May 23.
2
The language network as a natural kind within the broader landscape of the human brain.语言网络作为人类大脑更广阔景观中的一种自然类别。
Nat Rev Neurosci. 2024 May;25(5):289-312. doi: 10.1038/s41583-024-00802-4. Epub 2024 Apr 12.
3
Finding structure during incremental speech comprehension.
在逐步理解语音时寻找结构。
Elife. 2024 Apr 5;12:RP89311. doi: 10.7554/eLife.89311.
4
Emergent analogical reasoning in large language models.大语言模型中的紧急类比推理。
Nat Hum Behav. 2023 Sep;7(9):1526-1541. doi: 10.1038/s41562-023-01659-w. Epub 2023 Jul 31.
5
Modeling Structure-Building in the Brain With CCG Parsing and Large Language Models.用 CCG 解析和大型语言模型构建大脑的结构模型。
Cogn Sci. 2023 Jul;47(7):e13312. doi: 10.1111/cogs.13312.
6
Using cognitive psychology to understand GPT-3.利用认知心理学理解 GPT-3。
Proc Natl Acad Sci U S A. 2023 Feb 7;120(6):e2218523120. doi: 10.1073/pnas.2218523120. Epub 2023 Feb 2.
7
Shared computational principles for language processing in humans and deep language models.人类和深度语言模型语言处理的共享计算原则。
Nat Neurosci. 2022 Mar;25(3):369-380. doi: 10.1038/s41593-022-01026-4. Epub 2022 Mar 7.
8
The neural architecture of language: Integrative modeling converges on predictive processing.语言的神经结构:综合建模趋向于预测处理。
Proc Natl Acad Sci U S A. 2021 Nov 9;118(45). doi: 10.1073/pnas.2105646118.
9
Array programming with NumPy.使用 NumPy 进行数组编程。
Nature. 2020 Sep;585(7825):357-362. doi: 10.1038/s41586-020-2649-2. Epub 2020 Sep 16.
10
Lack of selectivity for syntax relative to word meanings throughout the language network.在整个语言网络中,相对于词义而言,对句法缺乏选择性。
Cognition. 2020 Oct;203:104348. doi: 10.1016/j.cognition.2020.104348. Epub 2020 Jun 20.