• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

听觉皮层通过空间分布的活动来编码唇读信息。

Auditory cortex encodes lipreading information through spatially distributed activity.

机构信息

Department of Psychology, University of Michigan, Ann Arbor, MI 48109, USA.

Department of Neurology, University of Michigan, Ann Arbor, MI 48109, USA.

出版信息

Curr Biol. 2024 Sep 9;34(17):4021-4032.e5. doi: 10.1016/j.cub.2024.07.073. Epub 2024 Aug 16.

DOI:10.1016/j.cub.2024.07.073
PMID:39153482
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11387126/
Abstract

Watching a speaker's face improves speech perception accuracy. This benefit is enabled, in part, by implicit lipreading abilities present in the general population. While it is established that lipreading can alter the perception of a heard word, it is unknown how these visual signals are represented in the auditory system or how they interact with auditory speech representations. One influential, but untested, hypothesis is that visual speech modulates the population-coded representations of phonetic and phonemic features in the auditory system. This model is largely supported by data showing that silent lipreading evokes activity in the auditory cortex, but these activations could alternatively reflect general effects of arousal or attention or the encoding of non-linguistic features such as visual timing information. This gap limits our understanding of how vision supports speech perception. To test the hypothesis that the auditory system encodes visual speech information, we acquired functional magnetic resonance imaging (fMRI) data from healthy adults and intracranial recordings from electrodes implanted in patients with epilepsy during auditory and visual speech perception tasks. Across both datasets, linear classifiers successfully decoded the identity of silently lipread words using the spatial pattern of auditory cortex responses. Examining the time course of classification using intracranial recordings, lipread words were classified at earlier time points relative to heard words, suggesting a predictive mechanism for facilitating speech. These results support a model in which the auditory system combines the joint neural distributions evoked by heard and lipread words to generate a more precise estimate of what was said.

摘要

观看说话者的面部表情可以提高言语感知的准确性。这种好处部分是由于普通人群中存在的隐性唇读能力。虽然已经确定唇读可以改变听到的单词的感知,但尚不清楚这些视觉信号如何在听觉系统中表示,以及它们如何与听觉言语表示相互作用。一个有影响力但未经测试的假设是,视觉言语调节了听觉系统中语音和音位特征的群体编码表示。这个模型在很大程度上得到了数据的支持,这些数据表明,无声唇读会在听觉皮层中引发活动,但这些激活也可能反映出一般的唤醒或注意力效应,或者非语言特征的编码,如视觉时间信息。这种差距限制了我们对视觉如何支持言语感知的理解。为了检验听觉系统编码视觉言语信息的假设,我们在健康成年人的功能磁共振成像(fMRI)数据和癫痫患者植入电极的颅内记录中,进行了听觉和视觉言语感知任务。在两个数据集上,线性分类器都成功地使用听觉皮层反应的空间模式来解码无声唇读单词的身份。通过颅内记录检查分类的时间过程,唇读单词相对于听到的单词可以更早地进行分类,这表明存在一种促进言语的预测机制。这些结果支持了一种模型,即听觉系统结合了由听到和唇读单词引起的联合神经分布,以生成更准确的说话内容估计。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/550a/11387126/5802663b5783/nihms-2014280-f0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/550a/11387126/78a651ea21bb/nihms-2014280-f0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/550a/11387126/94598e921a2c/nihms-2014280-f0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/550a/11387126/cb469077967d/nihms-2014280-f0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/550a/11387126/fd59bc36bb71/nihms-2014280-f0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/550a/11387126/5802663b5783/nihms-2014280-f0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/550a/11387126/78a651ea21bb/nihms-2014280-f0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/550a/11387126/94598e921a2c/nihms-2014280-f0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/550a/11387126/cb469077967d/nihms-2014280-f0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/550a/11387126/fd59bc36bb71/nihms-2014280-f0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/550a/11387126/5802663b5783/nihms-2014280-f0006.jpg

相似文献

1
Auditory cortex encodes lipreading information through spatially distributed activity.听觉皮层通过空间分布的活动来编码唇读信息。
Curr Biol. 2024 Sep 9;34(17):4021-4032.e5. doi: 10.1016/j.cub.2024.07.073. Epub 2024 Aug 16.
2
Activation of auditory cortex during silent lipreading.默读唇语时听觉皮层的激活。
Science. 1997 Apr 25;276(5312):593-6. doi: 10.1126/science.276.5312.593.
3
Increased Connectivity among Sensory and Motor Regions during Visual and Audiovisual Speech Perception.在视觉和视听言语感知过程中,感觉和运动区域之间的连通性增加。
J Neurosci. 2022 Jan 19;42(3):435-442. doi: 10.1523/JNEUROSCI.0114-21.2021. Epub 2021 Nov 23.
4
Lipreading and covert speech production similarly modulate human auditory-cortex responses to pure tones.唇读和隐性言语产生同样调节人类听觉皮层对纯音的反应。
J Neurosci. 2010 Jan 27;30(4):1314-21. doi: 10.1523/JNEUROSCI.1950-09.2010.
5
Attention to visual speech gestures enhances hemodynamic activity in the left planum temporale.对视觉言语手势的关注会增强左侧颞平面的血流动力学活动。
Hum Brain Mapp. 2006 Jun;27(6):471-7. doi: 10.1002/hbm.20190.
6
Visual speech perception without primary auditory cortex activation.无初级听觉皮层激活的视觉言语感知
Neuroreport. 2002 Mar 4;13(3):311-5. doi: 10.1097/00001756-200203040-00013.
7
A representation of abstract linguistic categories in the visual system underlies successful lipreading.视觉系统中抽象语言类别的表示是成功唇读的基础。
Neuroimage. 2023 Nov 15;282:120391. doi: 10.1016/j.neuroimage.2023.120391. Epub 2023 Sep 25.
8
Lip-Reading Enables the Brain to Synthesize Auditory Features of Unknown Silent Speech.唇读使大脑能够合成未知静音语音的听觉特征。
J Neurosci. 2020 Jan 29;40(5):1053-1065. doi: 10.1523/JNEUROSCI.1101-19.2019. Epub 2019 Dec 30.
9
Differential Auditory and Visual Phase-Locking Are Observed during Audio-Visual Benefit and Silent Lip-Reading for Speech Perception.在听觉-视觉获益和默读唇语感知语音时,观察到听觉和视觉相位锁定的差异。
J Neurosci. 2022 Aug 3;42(31):6108-6120. doi: 10.1523/JNEUROSCI.2476-21.2022. Epub 2022 Jun 27.
10
Primary auditory cortex activation by visual speech: an fMRI study at 3 T.视觉言语引起的初级听觉皮层激活:一项3T的功能磁共振成像研究
Neuroreport. 2005 Feb 8;16(2):125-8. doi: 10.1097/00001756-200502080-00010.

引用本文的文献

1
Neural Speech Tracking during Selective Attention: A Spatially Realistic Audiovisual Study.选择性注意期间的神经语音追踪:一项空间逼真的视听研究。
eNeuro. 2025 Jun 24;12(6). doi: 10.1523/ENEURO.0132-24.2025. Print 2025 Jun.
2
Evidence for a Causal Dissociation of the McGurk Effect and Congruent Audiovisual Speech Perception via TMS to the Left pSTS.经左颞上沟重复经颅磁刺激对 McGurk 效应和一致视听言语感知的因果分离的证据。
Multisens Res. 2024 Aug 16;37(4-5):341-363. doi: 10.1163/22134808-bja10129.
3
The effect of gaze on EEG measures of multisensory integration in a cocktail party scenario.

本文引用的文献

1
Large-scale single-neuron speech sound encoding across the depth of human cortex.人类大脑皮层深度范围内的大规模单神经元语音编码
Nature. 2024 Feb;626(7999):593-602. doi: 10.1038/s41586-023-06839-2. Epub 2023 Dec 13.
2
A representation of abstract linguistic categories in the visual system underlies successful lipreading.视觉系统中抽象语言类别的表示是成功唇读的基础。
Neuroimage. 2023 Nov 15;282:120391. doi: 10.1016/j.neuroimage.2023.120391. Epub 2023 Sep 25.
3
MEG Activity in Visual and Auditory Cortices Represents Acoustic Speech-Related Information during Silent Lip Reading.
在鸡尾酒会场景中,注视对多感官整合的脑电图测量的影响。
Front Hum Neurosci. 2023 Dec 15;17:1283206. doi: 10.3389/fnhum.2023.1283206. eCollection 2023.
4
Evidence for a Causal Dissociation of the McGurk Effect and Congruent Audiovisual Speech Perception via TMS.通过经颅磁刺激(TMS)证明麦格克效应与一致的视听语音感知之间存在因果分离。
bioRxiv. 2023 Nov 27:2023.11.27.568892. doi: 10.1101/2023.11.27.568892.
5
The effect of gaze on EEG measures of multisensory integration in a cocktail party scenario.在鸡尾酒会场景中,注视对多感官整合的脑电图测量的影响。
bioRxiv. 2023 Aug 24:2023.08.23.554451. doi: 10.1101/2023.08.23.554451.
大脑磁图活动在安静默读时代表了视觉和听觉皮层中的与声学语音相关的信息。
eNeuro. 2022 Jun 27;9(3). doi: 10.1523/ENEURO.0209-22.2022. Print 2022 May-Jun.
4
Lip movements enhance speech representations and effective connectivity in auditory dorsal stream.嘴唇运动增强了听觉背侧流中的言语表征和有效连通性。
Neuroimage. 2022 Aug 15;257:119311. doi: 10.1016/j.neuroimage.2022.119311. Epub 2022 May 16.
5
Visual speech differentially modulates beta, theta, and high gamma bands in auditory cortex.视觉言语对听觉皮层的 beta、theta 和高 gamma 频段有不同的调制作用。
Eur J Neurosci. 2021 Nov;54(9):7301-7317. doi: 10.1111/ejn.15482. Epub 2021 Oct 22.
6
Neurophysiological Indices of Audiovisual Speech Processing Reveal a Hierarchy of Multisensory Integration Effects.神经生理指标揭示视听言语加工中的多感觉整合效应层次结构。
J Neurosci. 2021 Jun 9;41(23):4991-5003. doi: 10.1523/JNEUROSCI.0906-20.2021. Epub 2021 Apr 6.
7
Convergence of heteromodal lexical retrieval in the lateral prefrontal cortex.外侧前额叶皮质中异模态词汇检索的收敛。
Sci Rep. 2021 Mar 18;11(1):6305. doi: 10.1038/s41598-021-85802-5.
8
The phase of cortical oscillations determines the perceptual fate of visual cues in naturalistic audiovisual speech.皮质振荡的相位决定了自然视听语音中视觉线索的感知命运。
Sci Adv. 2020 Nov 4;6(45). doi: 10.1126/sciadv.abc6348. Print 2020 Nov.
9
Crossmodal Phase Reset and Evoked Responses Provide Complementary Mechanisms for the Influence of Visual Speech in Auditory Cortex.跨模态相位重置和诱发反应为视觉语音在听觉皮层中的影响提供了互补的机制。
J Neurosci. 2020 Oct 28;40(44):8530-8542. doi: 10.1523/JNEUROSCI.0555-20.2020. Epub 2020 Oct 6.
10
Vision perceptually restores auditory spectral dynamics in speech.视觉在感知上恢复了言语中的听觉频谱动态。
Proc Natl Acad Sci U S A. 2020 Jul 21;117(29):16920-16927. doi: 10.1073/pnas.2002887117. Epub 2020 Jul 6.