• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

深度神经网络和人类都受益于组合式语言结构。

Deep neural networks and humans both benefit from compositional language structure.

作者信息

Galke Lukas, Ram Yoav, Raviv Limor

机构信息

Department of Mathematics and Computer Science, University of Southern Denmark, Odense, Denmark.

LEADS group, Max Planck Institute for Psycholinguistics, Nijmegen, Netherlands.

出版信息

Nat Commun. 2024 Dec 30;15(1):10816. doi: 10.1038/s41467-024-55158-1.

DOI:10.1038/s41467-024-55158-1
PMID:39738033
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11685529/
Abstract

Deep neural networks drive the success of natural language processing. A fundamental property of language is its compositional structure, allowing humans to systematically produce forms for new meanings. For humans, languages with more compositional and transparent structures are typically easier to learn than those with opaque and irregular structures. However, this learnability advantage has not yet been shown for deep neural networks, limiting their use as models for human language learning. Here, we directly test how neural networks compare to humans in learning and generalizing different languages that vary in their degree of compositional structure. We evaluate the memorization and generalization capabilities of a large language model and recurrent neural networks, and show that both deep neural networks exhibit a learnability advantage for more structured linguistic input: neural networks exposed to more compositional languages show more systematic generalization, greater agreement between different agents, and greater similarity to human learners.

摘要

深度神经网络推动了自然语言处理的成功。语言的一个基本特性是其组合结构,这使人类能够系统地为新含义生成形式。对于人类来说,具有更多组合性和透明结构的语言通常比具有不透明和不规则结构的语言更容易学习。然而,深度神经网络尚未展现出这种可学习性优势,这限制了它们作为人类语言学习模型的应用。在此,我们直接测试神经网络在学习和泛化不同组合结构程度的语言方面与人类相比情况如何。我们评估了一个大语言模型和循环神经网络的记忆与泛化能力,并表明这两种深度神经网络在处理结构更丰富的语言输入时都展现出可学习性优势:接触更多组合性语言的神经网络表现出更系统的泛化能力、不同智能体之间更高的一致性以及与人类学习者更高的相似性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8595/11685529/380a25dde75a/41467_2024_55158_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8595/11685529/0c0f78a462f7/41467_2024_55158_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8595/11685529/57a49119d9ab/41467_2024_55158_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8595/11685529/a7efd523eac3/41467_2024_55158_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8595/11685529/1b7a385a4a85/41467_2024_55158_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8595/11685529/380a25dde75a/41467_2024_55158_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8595/11685529/0c0f78a462f7/41467_2024_55158_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8595/11685529/57a49119d9ab/41467_2024_55158_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8595/11685529/a7efd523eac3/41467_2024_55158_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8595/11685529/1b7a385a4a85/41467_2024_55158_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8595/11685529/380a25dde75a/41467_2024_55158_Fig5_HTML.jpg

相似文献

1
Deep neural networks and humans both benefit from compositional language structure.深度神经网络和人类都受益于组合式语言结构。
Nat Commun. 2024 Dec 30;15(1):10816. doi: 10.1038/s41467-024-55158-1.
2
What makes a language easy to learn? A preregistered study on how systematic structure and community size affect language learnability.什么样的语言更容易学?一项关于系统性结构和社群规模如何影响语言可学度的预先注册研究。
Cognition. 2021 May;210:104620. doi: 10.1016/j.cognition.2021.104620. Epub 2021 Feb 8.
3
Linguistic generalization and compositionality in modern artificial neural networks.现代人工神经网络中的语言泛化和组合性。
Philos Trans R Soc Lond B Biol Sci. 2020 Feb 3;375(1791):20190307. doi: 10.1098/rstb.2019.0307. Epub 2019 Dec 16.
4
Finding Structure in One Child's Linguistic Experience.发现一个孩子的语言体验中的结构。
Cogn Sci. 2023 Jun;47(6):e13305. doi: 10.1111/cogs.13305.
5
Languages with more speakers tend to be harder to (machine-)learn.使用人口较多的语言往往更难(通过机器)学习。
Sci Rep. 2023 Oct 28;13(1):18521. doi: 10.1038/s41598-023-45373-z.
6
A Survey of the Usages of Deep Learning for Natural Language Processing.深度学习在自然语言处理中的应用调查。
IEEE Trans Neural Netw Learn Syst. 2021 Feb;32(2):604-624. doi: 10.1109/TNNLS.2020.2979670. Epub 2021 Feb 4.
7
Natural language instructions induce compositional generalization in networks of neurons.自然语言指令诱导神经元网络的组合泛化。
Nat Neurosci. 2024 May;27(5):988-999. doi: 10.1038/s41593-024-01607-5. Epub 2024 Mar 18.
8
Deep Learnability: Using Neural Networks to Quantify Language Similarity and Learnability.深度可学习性:使用神经网络量化语言相似度和可学习性。
Front Artif Intell. 2020 Jun 24;3:43. doi: 10.3389/frai.2020.00043. eCollection 2020.
9
Quasi-compositional mapping from form to meaning: a neural network-based approach to capturing neural responses during human language comprehension.从形式到意义的准组合映射:一种基于神经网络的方法,用于捕获人类语言理解过程中的神经反应。
Philos Trans R Soc Lond B Biol Sci. 2020 Feb 3;375(1791):20190313. doi: 10.1098/rstb.2019.0313. Epub 2019 Dec 16.
10
Human-like systematic generalization through a meta-learning neural network.通过元学习神经网络实现类人系统泛化。
Nature. 2023 Nov;623(7985):115-121. doi: 10.1038/s41586-023-06668-3. Epub 2023 Oct 25.

引用本文的文献

1
Effectiveness of the GPT-4o Model in Interpreting Electrocardiogram Images for Cardiac Diagnostics: Diagnostic Accuracy Study.GPT-4o模型在解读心电图图像用于心脏诊断中的有效性:诊断准确性研究
JMIR AI. 2025 Aug 22;4:e74426. doi: 10.2196/74426.

本文引用的文献

1
Language models, like humans, show content effects on reasoning tasks.语言模型和人类一样,在推理任务中表现出内容效应。
PNAS Nexus. 2024 Jul 16;3(7):pgae233. doi: 10.1093/pnasnexus/pgae233. eCollection 2024 Jul.
2
Grounded language acquisition through the eyes and ears of a single child.通过单个儿童的眼睛和耳朵进行基础语言习得。
Science. 2024 Feb 2;383(6682):504-511. doi: 10.1126/science.adi1374. Epub 2024 Feb 1.
3
Human-like systematic generalization through a meta-learning neural network.通过元学习神经网络实现类人系统泛化。
Nature. 2023 Nov;623(7985):115-121. doi: 10.1038/s41586-023-06668-3. Epub 2023 Oct 25.
4
Degrees of algorithmic equivalence between the brain and its DNN models.大脑与其 DNN 模型之间的算法等价程度。
Trends Cogn Sci. 2022 Dec;26(12):1090-1102. doi: 10.1016/j.tics.2022.09.003. Epub 2022 Oct 7.
5
Cumulative cultural evolution, population structure and the origin of combinatoriality in human language.累积文化进化、人口结构与人类语言组合性的起源。
Philos Trans R Soc Lond B Biol Sci. 2022 Jan 31;377(1843):20200319. doi: 10.1098/rstb.2020.0319. Epub 2021 Dec 13.
6
The neural architecture of language: Integrative modeling converges on predictive processing.语言的神经结构:综合建模趋向于预测处理。
Proc Natl Acad Sci U S A. 2021 Nov 9;118(45). doi: 10.1073/pnas.2105646118.
7
What makes a language easy to learn? A preregistered study on how systematic structure and community size affect language learnability.什么样的语言更容易学?一项关于系统性结构和社群规模如何影响语言可学度的预先注册研究。
Cognition. 2021 May;210:104620. doi: 10.1016/j.cognition.2021.104620. Epub 2021 Feb 8.
8
Performance vs. competence in human-machine comparisons.人机比较中的表现与能力。
Proc Natl Acad Sci U S A. 2020 Oct 27;117(43):26562-26571. doi: 10.1073/pnas.1905334117. Epub 2020 Oct 13.
9
Simplicity and informativeness in semantic category systems.语义类别系统的简洁性和信息量。
Cognition. 2020 Sep;202:104289. doi: 10.1016/j.cognition.2020.104289. Epub 2020 Jun 5.
10
Linguistic generalization and compositionality in modern artificial neural networks.现代人工神经网络中的语言泛化和组合性。
Philos Trans R Soc Lond B Biol Sci. 2020 Feb 3;375(1791):20190307. doi: 10.1098/rstb.2019.0307. Epub 2019 Dec 16.