• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

神经层面视听言语感知中熟悉度和一致性的自上而下预测

Top-Down Predictions of Familiarity and Congruency in Audio-Visual Speech Perception at Neural Level.

作者信息

Kolozsvári Orsolya B, Xu Weiyong, Leppänen Paavo H T, Hämäläinen Jarmo A

机构信息

Department of Psychology, University of Jyväskylä, Jyväskylä, Finland.

Jyväskylä Centre for Interdisciplinary Brain Research (CIBR), University of Jyväskylä, Jyväskylä, Finland.

出版信息

Front Hum Neurosci. 2019 Jul 12;13:243. doi: 10.3389/fnhum.2019.00243. eCollection 2019.

DOI:10.3389/fnhum.2019.00243
PMID:31354459
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC6639789/
Abstract

During speech perception, listeners rely on multimodal input and make use of both auditory and visual information. When presented with speech, for example syllables, the differences in brain responses to distinct stimuli are not, however, caused merely by the acoustic or visual features of the stimuli. The congruency of the auditory and visual information and the familiarity of a syllable, that is, whether it appears in the listener's native language or not, also modulates brain responses. We investigated how the congruency and familiarity of the presented stimuli affect brain responses to audio-visual (AV) speech in 12 adult Finnish native speakers and 12 adult Chinese native speakers. They watched videos of a Chinese speaker pronouncing syllables (/pa/, /pha/, /ta/, /tha/, /fa/) during a magnetoencephalography (MEG) measurement where only /pa/ and /ta/ were part of Finnish phonology while all the stimuli were part of Chinese phonology. The stimuli were presented in audio-visual (congruent or incongruent), audio only, or visual only conditions. The brain responses were examined in five time-windows: 75-125, 150-200, 200-300, 300-400, and 400-600 ms. We found significant differences for the congruency comparison in the fourth time-window (300-400 ms) in both sensor and source level analysis. Larger responses were observed for the incongruent stimuli than for the congruent stimuli. For the familiarity comparisons no significant differences were found. The results are in line with earlier studies reporting on the modulation of brain responses for audio-visual congruency around 250-500 ms. This suggests a much stronger process for the general detection of a mismatch between predictions based on lip movements and the auditory signal than for the top-down modulation of brain responses based on phonological information.

摘要

在言语感知过程中,听众依赖多模态输入并利用听觉和视觉信息。例如,当呈现语音(如音节)时,大脑对不同刺激的反应差异并非仅仅由刺激的声学或视觉特征引起。听觉和视觉信息的一致性以及音节的熟悉程度,即它是否出现在听众的母语中,也会调节大脑反应。我们研究了呈现刺激的一致性和熟悉程度如何影响12名成年芬兰母语者和12名成年中国母语者对视听(AV)语音的大脑反应。他们在脑磁图(MEG)测量期间观看了一位中国说话者发音节(/pa/、/pha/、/ta/、/tha/、/fa/)的视频,其中只有/pa/和/ta/是芬兰语音系的一部分,而所有刺激都是汉语音系的一部分。刺激以视听(一致或不一致)、仅听觉或仅视觉条件呈现。在五个时间窗口中检查大脑反应:75 - 125、150 - 200、200 - 300、300 - 400和400 - 600毫秒。我们发现在传感器和源水平分析的第四个时间窗口(300 - 400毫秒)中,一致性比较存在显著差异。观察到不一致刺激的反应比一致刺激的反应更大。对于熟悉程度比较,未发现显著差异。结果与早期关于250 - 500毫秒左右视听一致性对大脑反应调节的研究一致。这表明,与基于语音信息对大脑反应的自上而下调节相比,基于唇动预测和听觉信号之间不匹配的一般检测过程要强得多。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f555/6639789/de472623db06/fnhum-13-00243-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f555/6639789/48c6f81525c7/fnhum-13-00243-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f555/6639789/e101384a399d/fnhum-13-00243-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f555/6639789/44d0a500f520/fnhum-13-00243-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f555/6639789/de472623db06/fnhum-13-00243-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f555/6639789/48c6f81525c7/fnhum-13-00243-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f555/6639789/e101384a399d/fnhum-13-00243-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f555/6639789/44d0a500f520/fnhum-13-00243-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f555/6639789/de472623db06/fnhum-13-00243-g004.jpg

相似文献

1
Top-Down Predictions of Familiarity and Congruency in Audio-Visual Speech Perception at Neural Level.神经层面视听言语感知中熟悉度和一致性的自上而下预测
Front Hum Neurosci. 2019 Jul 12;13:243. doi: 10.3389/fnhum.2019.00243. eCollection 2019.
2
Perception of Incongruent Audiovisual Speech: Distribution of Modality-Specific Responses.对不一致的视听言语的感知:模态特异性反应的分布。
Am J Audiol. 2021 Dec 9;30(4):968-979. doi: 10.1044/2021_AJA-20-00213. Epub 2021 Sep 9.
3
Congruent audiovisual speech enhances auditory attention decoding with EEG.视听语音一致增强了 EEG 对听觉注意力的解码。
J Neural Eng. 2019 Nov 6;16(6):066033. doi: 10.1088/1741-2552/ab4340.
4
Audio-visual congruency alters power and coherence of oscillatory activity within and between cortical areas.视听一致性改变了皮质区域内和区域之间的振荡活动的功率和相干性。
Neuroimage. 2013 Oct 1;79:111-20. doi: 10.1016/j.neuroimage.2013.04.064. Epub 2013 May 2.
5
Sequential audiovisual interactions during speech perception: a whole-head MEG study.言语感知过程中的序列视听交互作用:一项全脑磁图研究。
Neuropsychologia. 2007 Mar 25;45(6):1342-54. doi: 10.1016/j.neuropsychologia.2006.09.019. Epub 2006 Oct 25.
6
Perception of incongruent audiovisual English consonants.感知不一致的英语视听辅音。
PLoS One. 2019 Mar 21;14(3):e0213588. doi: 10.1371/journal.pone.0213588. eCollection 2019.
7
Large Scale Functional Brain Networks Underlying Temporal Integration of Audio-Visual Speech Perception: An EEG Study.视听言语感知时间整合背后的大规模功能性脑网络:一项脑电图研究
Front Psychol. 2016 Oct 13;7:1558. doi: 10.3389/fpsyg.2016.01558. eCollection 2016.
8
Audio-visual onset differences are used to determine syllable identity for ambiguous audio-visual stimulus pairs.视听起始时间差异被用于确定视听刺激对中模糊的音节身份。
Front Psychol. 2013 Jun 26;4:331. doi: 10.3389/fpsyg.2013.00331. eCollection 2013.
9
Effects of congruent and incongruent visual cues on speech perception and brain activity in cochlear implant users.一致和不一致视觉线索对人工耳蜗使用者言语感知及大脑活动的影响。
Brain Struct Funct. 2015 Mar;220(2):1109-25. doi: 10.1007/s00429-013-0704-6. Epub 2014 Jan 9.
10
Effects of audio-visual integration on the detection of masked speech and non-speech sounds.视听整合对掩蔽语音和非语音声音检测的影响。
Brain Cogn. 2011 Feb;75(1):60-6. doi: 10.1016/j.bandc.2010.09.005. Epub 2010 Nov 9.

引用本文的文献

1
Interoceptive signals shape the earliest markers and neural pathway to awareness at the visual threshold.内感受信号塑造了视觉阈值处意识的最早标志物和神经通路。
Proc Natl Acad Sci U S A. 2024 Sep 10;121(37):e2311953121. doi: 10.1073/pnas.2311953121. Epub 2024 Sep 3.
2
Audio-visual combination of syllables involves time-sensitive dynamics following from fusion failure.音节的视听组合涉及融合失败后随时间敏感的动态变化。
Sci Rep. 2020 Oct 22;10(1):18009. doi: 10.1038/s41598-020-75201-7.

本文引用的文献

1
Audiovisual Processing of Chinese Characters Elicits Suppression and Congruency Effects in MEG.汉字的视听加工在脑磁图中引发抑制和一致性效应。
Front Hum Neurosci. 2019 Feb 6;13:18. doi: 10.3389/fnhum.2019.00018. eCollection 2019.
2
Neural pathways for visual speech perception.视觉言语感知的神经通路。
Front Neurosci. 2014 Dec 1;8:386. doi: 10.3389/fnins.2014.00386. eCollection 2014.
3
MEG and EEG data analysis with MNE-Python.使用 MNE-Python 进行 MEG 和 EEG 数据分析。
Front Neurosci. 2013 Dec 26;7:267. doi: 10.3389/fnins.2013.00267.
4
Electrophysiological evidence for speech-specific audiovisual integration.言语特异性视听整合的电生理学证据。
Neuropsychologia. 2014 Jan;53:115-21. doi: 10.1016/j.neuropsychologia.2013.11.011. Epub 2013 Nov 27.
5
FieldTrip: Open source software for advanced analysis of MEG, EEG, and invasive electrophysiological data.FieldTrip:用于 MEG、EEG 和有创电生理数据的高级分析的开源软件。
Comput Intell Neurosci. 2011;2011:156869. doi: 10.1155/2011/156869. Epub 2010 Dec 23.
6
fMRI-Guided transcranial magnetic stimulation reveals that the superior temporal sulcus is a cortical locus of the McGurk effect.功能磁共振成像引导经颅磁刺激显示,颞上沟是麦格克效应的皮质定位。
J Neurosci. 2010 Feb 17;30(7):2414-7. doi: 10.1523/JNEUROSCI.4865-09.2010.
7
Dual neural routing of visual facilitation in speech processing.言语处理中视觉促进的双重神经通路
J Neurosci. 2009 Oct 28;29(43):13445-53. doi: 10.1523/JNEUROSCI.3194-09.2009.
8
Visual anticipatory information modulates multisensory interactions of artificial audiovisual stimuli.视觉预期信息调节人工视听刺激的多感觉相互作用。
J Cogn Neurosci. 2010 Jul;22(7):1583-96. doi: 10.1162/jocn.2009.21308.
9
Neural signatures of phonetic learning in adulthood: a magnetoencephalography study.成人语音学习的神经特征:一项脑磁图研究。
Neuroimage. 2009 May 15;46(1):226-40. doi: 10.1016/j.neuroimage.2009.01.028. Epub 2009 Jan 29.
10
The mismatch negativity (MMN) in basic research of central auditory processing: a review.中枢听觉处理基础研究中的失配负波(MMN):综述
Clin Neurophysiol. 2007 Dec;118(12):2544-90. doi: 10.1016/j.clinph.2007.04.026. Epub 2007 Oct 10.