• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

卷积网络可以对与视觉单词识别过程中的前馈过程相关的脑磁图反应的功能调制进行建模。

Convolutional networks can model the functional modulation of the MEG responses associated with feed-forward processes during visual word recognition.

作者信息

van Vliet Marijn, Rinkinen Oona, Shimizu Takao, Niskanen Anni-Mari, Devereux Barry, Salmelin Riitta

机构信息

Department of Neuroscience and Biomedical Engineering, Aalto University, Espoo, Finland.

School of Electronics, Electrical Engineering and Computer Science, Queen's University Belfast, Belfast, United Kingdom.

出版信息

Elife. 2025 May 13;13:RP96217. doi: 10.7554/eLife.96217.

DOI:10.7554/eLife.96217
PMID:40359126
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12074637/
Abstract

Traditional models of reading lack a realistic simulation of the early visual processing stages, taking input in the form of letter banks and predefined line segments, making them unsuitable for modeling early brain responses. We used variations of the VGG-11 convolutional neural network (CNN) to create models of visual word recognition that starts from the pixel-level and performs the macro-scale computations needed for the detection and segmentation of letter shapes to word-form identification of large vocabulary of 10k Finnish words, regardless of letter size, shape, or rotation. The models were evaluated based on an existing magnetoencephalography (MEG) study where participants viewed regular words, pseudowords, noise-embedded words, symbol strings, and consonant strings. The original images used in the study were presented to the models and the activity in the layers was compared to MEG evoked response amplitudes. Through a few alterations to make the network more biologically plausible, we found an CNN architecture that can correctly simulate the behavior of three prominent responses, namely the type I (early visual response), type II (the 'letter string' response), and the N400m. In conclusion, starting a model of reading with convolution-and-pooling steps enables the flexibility and realism crucial for a direct model-to-brain comparison.

摘要

传统的阅读模型缺乏对早期视觉处理阶段的真实模拟,它们以字母库和预定义线段的形式获取输入,因此不适用于对早期大脑反应进行建模。我们使用VGG - 11卷积神经网络(CNN)的变体来创建视觉单词识别模型,该模型从像素级别开始,执行从字母形状检测和分割到10000个芬兰语单词的大词汇量单词形式识别所需的宏观尺度计算,而不考虑字母的大小、形状或旋转。这些模型是基于一项现有的脑磁图(MEG)研究进行评估的,在该研究中,参与者观看了常规单词、伪单词、嵌入噪声的单词、符号串和辅音串。将研究中使用的原始图像呈现给模型,并将各层的活动与MEG诱发反应幅度进行比较。通过一些调整使网络在生物学上更合理,我们发现了一种CNN架构,它可以正确模拟三种突出反应的行为,即I型(早期视觉反应)、II型(“字母串”反应)和N400m。总之,从卷积和池化步骤开始构建阅读模型能够实现对大脑进行直接模型比较至关重要的灵活性和真实性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/4818962e8d0a/elife-96217-fig7.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/760a0b261ec3/elife-96217-fig1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/5271b7700671/elife-96217-fig1-figsupp1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/ebbf80bc19cb/elife-96217-fig2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/944ffc7cf5ed/elife-96217-fig3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/f5df8269e1fe/elife-96217-fig3-figsupp1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/19cb12ab0c0a/elife-96217-fig3-figsupp2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/459974423f11/elife-96217-fig3-figsupp3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/f97537140369/elife-96217-fig4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/8302673f2130/elife-96217-fig5.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/62c540cfdd3c/elife-96217-fig5-figsupp1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/78248dad3f6e/elife-96217-fig5-figsupp2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/5f3993ae1c6c/elife-96217-fig5-figsupp3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/5adb10484275/elife-96217-fig5-figsupp4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/699b6df42707/elife-96217-fig6.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/dd8e9267d9d6/elife-96217-fig6-figsupp1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/4818962e8d0a/elife-96217-fig7.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/760a0b261ec3/elife-96217-fig1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/5271b7700671/elife-96217-fig1-figsupp1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/ebbf80bc19cb/elife-96217-fig2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/944ffc7cf5ed/elife-96217-fig3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/f5df8269e1fe/elife-96217-fig3-figsupp1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/19cb12ab0c0a/elife-96217-fig3-figsupp2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/459974423f11/elife-96217-fig3-figsupp3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/f97537140369/elife-96217-fig4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/8302673f2130/elife-96217-fig5.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/62c540cfdd3c/elife-96217-fig5-figsupp1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/78248dad3f6e/elife-96217-fig5-figsupp2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/5f3993ae1c6c/elife-96217-fig5-figsupp3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/5adb10484275/elife-96217-fig5-figsupp4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/699b6df42707/elife-96217-fig6.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/dd8e9267d9d6/elife-96217-fig6-figsupp1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a7ba/12074637/4818962e8d0a/elife-96217-fig7.jpg

相似文献

1
Convolutional networks can model the functional modulation of the MEG responses associated with feed-forward processes during visual word recognition.卷积网络可以对与视觉单词识别过程中的前馈过程相关的脑磁图反应的功能调制进行建模。
Elife. 2025 May 13;13:RP96217. doi: 10.7554/eLife.96217.
2
A compositional neural code in high-level visual cortex can explain jumbled word reading.高级视觉皮层中的组合神经代码可以解释乱序单词阅读。
Elife. 2020 May 5;9:e54846. doi: 10.7554/eLife.54846.
3
Word-specific repetition effects revealed by MEG and the implications for lexical access.基于 MEG 的词特异性重复效应及其对词汇通达的启示
Brain Lang. 2013 Dec;127(3):497-509. doi: 10.1016/j.bandl.2013.09.013. Epub 2013 Oct 29.
4
Differential Phonological and Semantic Modulation of Neurophysiological Responses to Visual Word Recognition.对视觉单词识别的神经生理反应的差异音韵和语义调制
Neuropsychobiology. 2015;72(1):46-56. doi: 10.1159/000379752.
5
Functional magnetic resonance imaging blood oxygenation level-dependent signal and magnetoencephalography evoked responses yield different neural functionality in reading.功能磁共振成像血氧水平依赖信号和脑磁图诱发反应在阅读中产生不同的神经功能。
J Neurosci. 2011 Jan 19;31(3):1048-58. doi: 10.1523/JNEUROSCI.3113-10.2011.
6
From orthography to phonetics: ERP measures of grapheme-to-phoneme conversion mechanisms in reading.从正字法到语音学:阅读中字形到音素转换机制的事件相关电位测量
J Cogn Neurosci. 2004 Mar;16(2):301-17. doi: 10.1162/089892904322984580.
7
Information properties of morphologically complex words modulate brain activity during word reading.形态复杂词的信息属性在单词阅读过程中调节大脑活动。
Hum Brain Mapp. 2018 Jun;39(6):2583-2595. doi: 10.1002/hbm.24025. Epub 2018 Mar 9.
8
The consonant/vowel pattern determines the structure of orthographic representations in the left fusiform gyrus.辅音/元音模式决定了左梭状回中拼字法表现的结构。
Cortex. 2018 Apr;101:73-86. doi: 10.1016/j.cortex.2018.01.006. Epub 2018 Feb 2.
9
Cracking the neural code for word recognition in convolutional neural networks.破解卷积神经网络中单词识别的神经密码。
PLoS Comput Biol. 2024 Sep 6;20(9):e1012430. doi: 10.1371/journal.pcbi.1012430. eCollection 2024 Sep.
10
Predictive pre-activation of orthographic and lexical-semantic representations facilitates visual word recognition.预测性预激活字形和词汇语义表征可促进视觉词汇识别。
Psychophysiology. 2022 Mar;59(3):e13970. doi: 10.1111/psyp.13970. Epub 2021 Nov 23.

本文引用的文献

1
A predictive coding model of the N400.N400的预测编码模型。
Cognition. 2024 May;246:105755. doi: 10.1016/j.cognition.2024.105755. Epub 2024 Feb 29.
2
The neuroconnectionist research programme.神经连接主义研究计划。
Nat Rev Neurosci. 2023 Jul;24(7):431-450. doi: 10.1038/s41583-023-00705-w. Epub 2023 May 30.
3
Deep problems with neural network models of human vision.人类视觉神经网络模型的深层问题。
Behav Brain Sci. 2022 Dec 1;46:e385. doi: 10.1017/S0140525X22002813.
4
A separable neural code in monkey IT enables perfect CAPTCHA decoding.猴子 IT 中的可分离神经代码可实现完美的 CAPTCHA 解码。
J Neurophysiol. 2022 Apr 1;127(4):869-884. doi: 10.1152/jn.00160.2021. Epub 2022 Feb 23.
5
Brains and algorithms partially converge in natural language processing.大脑和算法在自然语言处理中部分融合。
Commun Biol. 2022 Feb 16;5(1):134. doi: 10.1038/s42003-022-03036-1.
6
Emergence of a compositional neural code for written words: Recycling of a convolutional neural network for reading.书面文字组合神经编码的出现:用于阅读的卷积神经网络的再利用。
Proc Natl Acad Sci U S A. 2021 Nov 16;118(46). doi: 10.1073/pnas.2104779118.
7
Spatiotemporal dynamics of orthographic and lexical processing in the ventral visual pathway.腹侧视觉通路上的字形和词汇处理的时空动态。
Nat Hum Behav. 2021 Mar;5(3):389-398. doi: 10.1038/s41562-020-00982-w. Epub 2020 Nov 30.
8
The inferior temporal cortex is a potential cortical precursor of orthographic processing in untrained monkeys.下颞叶皮层是未经训练的猴子在字形处理中的潜在皮质前体。
Nat Commun. 2020 Aug 4;11(1):3886. doi: 10.1038/s41467-020-17714-3.
9
A compositional neural code in high-level visual cortex can explain jumbled word reading.高级视觉皮层中的组合神经代码可以解释乱序单词阅读。
Elife. 2020 May 5;9:e54846. doi: 10.7554/eLife.54846.
10
An orthographic prediction error as the basis for efficient visual word recognition.基于正字法预测错误的高效视觉词识别。
Neuroimage. 2020 Jul 1;214:116727. doi: 10.1016/j.neuroimage.2020.116727. Epub 2020 Mar 12.