• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

通过多模态信息整合实现数感的出现:来自神经网络模型的发展性学习见解

Emergence of number sense through the integration of multimodal information: developmental learning insights from neural network models.

作者信息

Noda Kamma, Soda Takafumi, Yamashita Yuichi

机构信息

Department of Information Medicine, National Institute of Neuroscience, National Center of Neurology and Psychiatry, Kodaira, Japan.

出版信息

Front Neurosci. 2024 Jan 17;18:1330512. doi: 10.3389/fnins.2024.1330512. eCollection 2024.

DOI:10.3389/fnins.2024.1330512
PMID:38298912
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10828047/
Abstract

INTRODUCTION

Associating multimodal information is essential for human cognitive abilities including mathematical skills. Multimodal learning has also attracted attention in the field of machine learning, and it has been suggested that the acquisition of better latent representation plays an important role in enhancing task performance. This study aimed to explore the impact of multimodal learning on representation, and to understand the relationship between multimodal representation and the development of mathematical skills.

METHODS

We employed a multimodal deep neural network as the computational model for multimodal associations in the brain. We compared the representations of numerical information, that is, handwritten digits and images containing a variable number of geometric figures learned through single- and multimodal methods. Next, we evaluated whether these representations were beneficial for downstream arithmetic tasks.

RESULTS

Multimodal training produced better latent representation in terms of clustering quality, which is consistent with previous findings on multimodal learning in deep neural networks. Moreover, the representations learned using multimodal information exhibited superior performance in arithmetic tasks.

DISCUSSION

Our novel findings experimentally demonstrate that changes in acquired latent representations through multimodal association learning are directly related to cognitive functions, including mathematical skills. This supports the possibility that multimodal learning using deep neural network models may offer novel insights into higher cognitive functions.

摘要

引言

关联多模态信息对于包括数学技能在内的人类认知能力至关重要。多模态学习在机器学习领域也引起了关注,并且有人提出,获得更好的潜在表征在提高任务性能方面起着重要作用。本研究旨在探讨多模态学习对表征的影响,并了解多模态表征与数学技能发展之间的关系。

方法

我们采用多模态深度神经网络作为大脑中多模态关联的计算模型。我们比较了通过单模态和多模态方法学习的数字信息(即手写数字和包含不同数量几何图形的图像)的表征。接下来,我们评估了这些表征是否对下游算术任务有益。

结果

就聚类质量而言,多模态训练产生了更好的潜在表征,这与先前关于深度神经网络中多模态学习的研究结果一致。此外,使用多模态信息学习的表征在算术任务中表现出卓越的性能。

讨论

我们的新发现通过实验证明,通过多模态关联学习获得的潜在表征的变化与包括数学技能在内的认知功能直接相关。这支持了使用深度神经网络模型的多模态学习可能为更高层次的认知功能提供新见解的可能性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae3d/10828047/f6a8987c582e/fnins-18-1330512-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae3d/10828047/68a8e4b5128d/fnins-18-1330512-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae3d/10828047/2ebd186950a8/fnins-18-1330512-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae3d/10828047/d274660b5e7f/fnins-18-1330512-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae3d/10828047/be728db7f226/fnins-18-1330512-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae3d/10828047/f5eb5a862779/fnins-18-1330512-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae3d/10828047/1a3f4b9edd35/fnins-18-1330512-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae3d/10828047/78faca3c05a2/fnins-18-1330512-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae3d/10828047/283c04cc300f/fnins-18-1330512-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae3d/10828047/34f4ebd5b5f6/fnins-18-1330512-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae3d/10828047/f63acc797ca8/fnins-18-1330512-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae3d/10828047/4565fca17360/fnins-18-1330512-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae3d/10828047/f6a8987c582e/fnins-18-1330512-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae3d/10828047/68a8e4b5128d/fnins-18-1330512-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae3d/10828047/2ebd186950a8/fnins-18-1330512-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae3d/10828047/d274660b5e7f/fnins-18-1330512-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae3d/10828047/be728db7f226/fnins-18-1330512-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae3d/10828047/f5eb5a862779/fnins-18-1330512-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae3d/10828047/1a3f4b9edd35/fnins-18-1330512-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae3d/10828047/78faca3c05a2/fnins-18-1330512-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae3d/10828047/283c04cc300f/fnins-18-1330512-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae3d/10828047/34f4ebd5b5f6/fnins-18-1330512-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae3d/10828047/f63acc797ca8/fnins-18-1330512-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae3d/10828047/4565fca17360/fnins-18-1330512-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ae3d/10828047/f6a8987c582e/fnins-18-1330512-g012.jpg

相似文献

1
Emergence of number sense through the integration of multimodal information: developmental learning insights from neural network models.通过多模态信息整合实现数感的出现:来自神经网络模型的发展性学习见解
Front Neurosci. 2024 Jan 17;18:1330512. doi: 10.3389/fnins.2024.1330512. eCollection 2024.
2
Performance of a Computational Model of the Mammalian Olfactory System哺乳动物嗅觉系统计算模型的性能
3
Development of novel deep multimodal representation learning-based model for the differentiation of liver tumors on B-mode ultrasound images.基于新型深度多模态表示学习的 B 型超声图像肝脏肿瘤鉴别模型的构建。
J Gastroenterol Hepatol. 2022 Apr;37(4):678-684. doi: 10.1111/jgh.15763. Epub 2022 Jan 12.
4
Multimodal information bottleneck for deep reinforcement learning with multiple sensors.多模态信息瓶颈用于多传感器的深度强化学习。
Neural Netw. 2024 Aug;176:106347. doi: 10.1016/j.neunet.2024.106347. Epub 2024 Apr 27.
5
Ki-Cook: clustering multimodal cooking representations through knowledge-infused learning.Ki-Cook:通过知识注入学习对多模态烹饪表示进行聚类
Front Big Data. 2023 Jul 24;6:1200840. doi: 10.3389/fdata.2023.1200840. eCollection 2023.
6
Numerosity discrimination in deep neural networks: Initial competence, developmental refinement and experience statistics.深度神经网络中的数量辨别:初始能力、发展细化和经验统计。
Dev Sci. 2020 Sep;23(5):e12940. doi: 10.1111/desc.12940. Epub 2020 Feb 18.
7
Frontoparietal and salience network synchronizations during nonsymbolic magnitude processing predict brain age and mathematical performance in youth.非符号数量加工过程中的额顶网络和突显网络同步可以预测青少年的大脑年龄和数学表现。
Hum Brain Mapp. 2024 Aug 1;45(11):e26777. doi: 10.1002/hbm.26777.
8
Associating Latent Representations With Cognitive Maps via Hyperspherical Space for Neural Population Spikes.通过超球面关联潜在表示与神经群体峰的认知图。
IEEE Trans Neural Syst Rehabil Eng. 2022;30:2886-2895. doi: 10.1109/TNSRE.2022.3212997. Epub 2022 Oct 20.
9
Harmonized Multimodal Learning with Gaussian Process Latent Variable Models.基于高斯过程潜变量模型的协调多模态学习。
IEEE Trans Pattern Anal Mach Intell. 2021 Mar;43(3):858-872. doi: 10.1109/TPAMI.2019.2942028. Epub 2021 Feb 4.
10
Learning Numerosity Representations with Transformers: Number Generation Tasks and Out-of-Distribution Generalization.使用Transformer学习数量表示:数字生成任务与分布外泛化
Entropy (Basel). 2021 Jul 3;23(7):857. doi: 10.3390/e23070857.

本文引用的文献

1
Comparison of visual quantities in untrained neural networks.未经训练的神经网络中的视觉量比较。
Cell Rep. 2023 Aug 29;42(8):112900. doi: 10.1016/j.celrep.2023.112900. Epub 2023 Jul 29.
2
Learning-induced reorganization of number neurons and emergence of numerical representations in a biologically inspired neural network.在一个受生物启发的神经网络中,学习引起的数字神经元的重组和数值表示的出现。
Nat Commun. 2023 Jun 29;14(1):3843. doi: 10.1038/s41467-023-39548-5.
3
Artificial neural network modelling of the neural population code underlying mathematical operations.
人工神经网络对数学运算所基于的神经群体编码的建模。
Neuroimage. 2023 Apr 15;270:119980. doi: 10.1016/j.neuroimage.2023.119980. Epub 2023 Feb 26.
4
A visual sense of number emerges from divisive normalization in a simple center-surround convolutional network.一种简单的中心环绕卷积网络中的除法归一化产生了对数字的视觉感知。
Elife. 2022 Oct 3;11:e80990. doi: 10.7554/eLife.80990.
5
Numerosity tuning in human association cortices and local image contrast representations in early visual cortex.人类联合皮层的数量调谐和早期视觉皮层中的局部图像对比表示。
Nat Commun. 2022 Mar 15;13(1):1340. doi: 10.1038/s41467-022-29030-z.
6
Spontaneous representation of numerosity zero in a deep neural network for visual object recognition.用于视觉目标识别的深度神经网络中数字零的自发表征。
iScience. 2021 Oct 15;24(11):103301. doi: 10.1016/j.isci.2021.103301. eCollection 2021 Nov 19.
7
Learning Numerosity Representations with Transformers: Number Generation Tasks and Out-of-Distribution Generalization.使用Transformer学习数量表示:数字生成任务与分布外泛化
Entropy (Basel). 2021 Jul 3;23(7):857. doi: 10.3390/e23070857.
8
Enhanced perception and memory: Insights from synesthesia and expertise.增强的感知和记忆:联觉和专业知识的启示。
Cortex. 2021 Jul;140:14-25. doi: 10.1016/j.cortex.2021.01.024. Epub 2021 Mar 31.
9
Paradoxical sensory reactivity induced by functional disconnection in a robot model of neurodevelopmental disorder.功能连接中断诱导神经发育障碍机器人模型中的矛盾感觉反应。
Neural Netw. 2021 Jun;138:150-163. doi: 10.1016/j.neunet.2021.01.033. Epub 2021 Feb 12.
10
Visual number sense in untrained deep neural networks.未训练的深度神经网络中的视觉数字感知。
Sci Adv. 2021 Jan 1;7(1). doi: 10.1126/sciadv.abd6127. Print 2021 Jan.