Suppr超能文献

听觉-视觉语音和非语音信号感知过程中的跨模态相互作用:一项 fMRI 研究。

Cross-modal interactions during perception of audiovisual speech and nonspeech signals: an fMRI study.

机构信息

Department of General Neurology, University of Tübingen, Tübingen, Germany.

出版信息

J Cogn Neurosci. 2011 Jan;23(1):221-37. doi: 10.1162/jocn.2010.21421.

Abstract

During speech communication, visual information may interact with the auditory system at various processing stages. Most noteworthy, recent magnetoencephalography (MEG) data provided first evidence for early and preattentive phonetic/phonological encoding of the visual data stream--prior to its fusion with auditory phonological features [Hertrich, I., Mathiak, K., Lutzenberger, W., & Ackermann, H. Time course of early audiovisual interactions during speech and non-speech central-auditory processing: An MEG study. Journal of Cognitive Neuroscience, 21, 259-274, 2009]. Using functional magnetic resonance imaging, the present follow-up study aims to further elucidate the topographic distribution of visual-phonological operations and audiovisual (AV) interactions during speech perception. Ambiguous acoustic syllables--disambiguated to /pa/ or /ta/ by the visual channel (speaking face)--served as test materials, concomitant with various control conditions (nonspeech AV signals, visual-only and acoustic-only speech, and nonspeech stimuli). (i) Visual speech yielded an AV-subadditive activation of primary auditory cortex and the anterior superior temporal gyrus (STG), whereas the posterior STG responded both to speech and nonspeech motion. (ii) The inferior frontal and the fusiform gyrus of the right hemisphere showed a strong phonetic/phonological impact (differential effects of visual /pa/ vs. /ta/) upon hemodynamic activation during presentation of speaking faces. Taken together with the previous MEG data, these results point at a dual-pathway model of visual speech information processing: On the one hand, access to the auditory system via the anterior supratemporal “what" path may give rise to direct activation of "auditory objects." On the other hand, visual speech information seems to be represented in a right-hemisphere visual working memory, providing a potential basis for later interactions with auditory information such as the McGurk effect.

摘要

在言语交流中,视觉信息可能会在各个处理阶段与听觉系统相互作用。最值得注意的是,最近的脑磁图(MEG)数据首次提供了证据,证明在视觉数据流与听觉语音特征融合之前,就已经对其进行了早期和非注意语音/音系编码[Hertrich,I.,Mathiak,K.,Lutzenberger,W.,&Ackermann,H. 语音和非语音中枢听觉处理过程中视听早期相互作用的时间进程:一项 MEG 研究。认知神经科学杂志,21,259-274,2009]。使用功能磁共振成像,本后续研究旨在进一步阐明视觉-语音操作和视听(AV)相互作用在言语感知中的拓扑分布。作为测试材料的是具有歧义的声学音节-通过视觉通道(说话的脸)分辨为/pa/或/ta/-伴随着各种对照条件(非语音 AV 信号、视觉仅和声学仅语音、非语音刺激)。(i)视觉语音导致初级听觉皮层和前上颞叶(STG)的 AV 亚加性激活,而 STG 后部则对语音和非语音运动都有反应。(ii)右半球的下额叶和梭状回表现出强烈的语音/音系影响(在呈现说话脸时,视觉/pa/与/ta/的差异效应),对血流动力学激活有影响。结合之前的 MEG 数据,这些结果表明视觉语音信息处理存在双通路模型:一方面,通过前上颞叶的“什么”路径进入听觉系统可能会导致“听觉对象”的直接激活。另一方面,视觉语音信息似乎在右半球的视觉工作记忆中得到表示,为后来与听觉信息的相互作用提供了潜在基础,例如麦格克效应。

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验