• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

皮质振荡的相位决定了自然视听语音中视觉线索的感知命运。

The phase of cortical oscillations determines the perceptual fate of visual cues in naturalistic audiovisual speech.

机构信息

Department of Basic Neurosciences, Faculty of Medicine, University of Geneva, 1202 Geneva, Switzerland.

Division of Neurology, Department of Clinical Neurosciences, Geneva University Hospitals, 1205 Geneva, Switzerland.

出版信息

Sci Adv. 2020 Nov 4;6(45). doi: 10.1126/sciadv.abc6348. Print 2020 Nov.

DOI:10.1126/sciadv.abc6348
PMID:33148648
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7673697/
Abstract

When we see our interlocutor, our brain seamlessly extracts visual cues from their face and processes them along with the sound of their voice, making speech an intrinsically multimodal signal. Visual cues are especially important in noisy environments, when the auditory signal is less reliable. Neuronal oscillations might be involved in the cortical processing of audiovisual speech by selecting which sensory channel contributes more to perception. To test this, we designed computer-generated naturalistic audiovisual speech stimuli where one mismatched phoneme-viseme pair in a key word of sentences created bistable perception. Neurophysiological recordings (high-density scalp and intracranial electroencephalography) revealed that the precise phase angle of theta-band oscillations in posterior temporal and occipital cortex of the right hemisphere was crucial to select whether the auditory or the visual speech cue drove perception. We demonstrate that the phase of cortical oscillations acts as an instrument for sensory selection in audiovisual speech processing.

摘要

当我们看到对话者时,大脑会从他们的面部无缝提取视觉线索,并与声音一起处理这些线索,使言语成为一种内在的多模态信号。在嘈杂的环境中,听觉信号不太可靠时,视觉线索尤为重要。神经元振荡可能参与了视听言语的皮层处理,选择哪个感觉通道对感知的贡献更大。为了验证这一点,我们设计了计算机生成的自然视听言语刺激,在句子的关键词中,一个不匹配的音位-视位对会产生双稳态感知。神经生理学记录(高密度头皮和颅内脑电图)显示,右半球后颞叶和枕叶皮质中θ带振荡的精确相位角对于选择听觉还是视觉言语线索驱动感知至关重要。我们证明,皮层振荡的相位在视听言语处理中充当了感觉选择的工具。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0a5/7673697/efadbc0ed7c1/abc6348-F5.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0a5/7673697/146d753d164b/abc6348-F1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0a5/7673697/611a20be37ea/abc6348-F2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0a5/7673697/3e00603a240e/abc6348-F3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0a5/7673697/d0929b241c1f/abc6348-F4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0a5/7673697/efadbc0ed7c1/abc6348-F5.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0a5/7673697/146d753d164b/abc6348-F1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0a5/7673697/611a20be37ea/abc6348-F2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0a5/7673697/3e00603a240e/abc6348-F3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0a5/7673697/d0929b241c1f/abc6348-F4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a0a5/7673697/efadbc0ed7c1/abc6348-F5.jpg

相似文献

1
The phase of cortical oscillations determines the perceptual fate of visual cues in naturalistic audiovisual speech.皮质振荡的相位决定了自然视听语音中视觉线索的感知命运。
Sci Adv. 2020 Nov 4;6(45). doi: 10.1126/sciadv.abc6348. Print 2020 Nov.
2
Speech-specific audiovisual integration modulates induced theta-band oscillations.语音特异性视听整合调制诱导的 theta 波段振荡。
PLoS One. 2019 Jul 16;14(7):e0219744. doi: 10.1371/journal.pone.0219744. eCollection 2019.
3
Left Motor δ Oscillations Reflect Asynchrony Detection in Multisensory Speech Perception.左运动 δ 振荡反映多感觉语音感知中的异步检测。
J Neurosci. 2022 Mar 16;42(11):2313-2326. doi: 10.1523/JNEUROSCI.2965-20.2022. Epub 2022 Jan 27.
4
Generalizable EEG Encoding Models with Naturalistic Audiovisual Stimuli.具有自然视听刺激的可泛化 EEG 编码模型。
J Neurosci. 2021 Oct 27;41(43):8946-8962. doi: 10.1523/JNEUROSCI.2891-20.2021. Epub 2021 Sep 9.
5
Effects of congruent and incongruent visual cues on speech perception and brain activity in cochlear implant users.一致和不一致视觉线索对人工耳蜗使用者言语感知及大脑活动的影响。
Brain Struct Funct. 2015 Mar;220(2):1109-25. doi: 10.1007/s00429-013-0704-6. Epub 2014 Jan 9.
6
Degradation of labial information modifies audiovisual speech perception in cochlear-implanted children.唇语信息的退化改变了植入人工耳蜗的儿童对视听语音的感知。
Ear Hear. 2013 Jan-Feb;34(1):110-21. doi: 10.1097/AUD.0b013e3182670993.
7
Prediction and constraint in audiovisual speech perception.视听言语感知中的预测与约束
Cortex. 2015 Jul;68:169-81. doi: 10.1016/j.cortex.2015.03.006. Epub 2015 Mar 20.
8
Deficits in audiovisual speech perception in normal aging emerge at the level of whole-word recognition.正常衰老过程中视听言语感知的缺陷出现在整词识别层面。
Neurobiol Aging. 2015 Jan;36(1):283-91. doi: 10.1016/j.neurobiolaging.2014.08.003. Epub 2014 Aug 7.
9
Eye Can Hear Clearly Now: Inverse Effectiveness in Natural Audiovisual Speech Processing Relies on Long-Term Crossmodal Temporal Integration.现在眼睛能“听清”了:自然视听言语处理中的反向有效性依赖于长期跨模态时间整合。
J Neurosci. 2016 Sep 21;36(38):9888-95. doi: 10.1523/JNEUROSCI.1396-16.2016.
10
Perception of audiovisual infant directed speech.感知视听婴儿导向语音。
Scand J Psychol. 2020 Apr;61(2):218-226. doi: 10.1111/sjop.12599. Epub 2019 Dec 9.

引用本文的文献

1
Impact of age-related hearing loss on decompensation of left DLPFC during speech perception in noise: a combined EEG-fNIRS study.年龄相关性听力损失对噪声环境下言语感知过程中左侧背外侧前额叶皮质失代偿的影响:一项脑电图-功能近红外光谱联合研究
Geroscience. 2025 Apr;47(2):2119-2134. doi: 10.1007/s11357-024-01393-9. Epub 2024 Oct 24.
2
Brain Encoding of Naturalistic, Continuous, and Unpredictable Tactile Events.自然连续不可预测触觉事件的大脑编码。
eNeuro. 2024 Sep 24;11(9). doi: 10.1523/ENEURO.0238-24.2024. Print 2024 Sep.
3
Auditory cortex encodes lipreading information through spatially distributed activity.

本文引用的文献

1
Crossmodal Phase Reset and Evoked Responses Provide Complementary Mechanisms for the Influence of Visual Speech in Auditory Cortex.跨模态相位重置和诱发反应为视觉语音在听觉皮层中的影响提供了互补的机制。
J Neurosci. 2020 Oct 28;40(44):8530-8542. doi: 10.1523/JNEUROSCI.0555-20.2020. Epub 2020 Oct 6.
2
Animated virtual characters to explore audio-visual speech in controlled and naturalistic environments.用于在受控和自然环境中探索视听语音的动画虚拟角色。
Sci Rep. 2020 Sep 23;10(1):15540. doi: 10.1038/s41598-020-72375-y.
3
The phase of pre-stimulus brain oscillations correlates with cross-modal synchrony perception.
听觉皮层通过空间分布的活动来编码唇读信息。
Curr Biol. 2024 Sep 9;34(17):4021-4032.e5. doi: 10.1016/j.cub.2024.07.073. Epub 2024 Aug 16.
4
Multi-timescale neural dynamics for multisensory integration.多感觉整合的多时间尺度神经动力学
Nat Rev Neurosci. 2024 Sep;25(9):625-642. doi: 10.1038/s41583-024-00845-7. Epub 2024 Aug 1.
5
Dynamic causal modeling analysis reveals the modulation of motor cortex and integration in superior temporal gyrus during multisensory speech perception.动态因果模型分析揭示了多感官语音感知过程中运动皮层的调节作用以及颞上回的整合作用。
Cogn Neurodyn. 2024 Jun;18(3):931-946. doi: 10.1007/s11571-023-09945-z. Epub 2023 Mar 4.
6
Phase-dependent word perception emerges from region-specific sensitivity to the statistics of language.相位相关的单词感知源于对语言统计数据的区域特异性敏感性。
Proc Natl Acad Sci U S A. 2024 Jun 4;121(23):e2320489121. doi: 10.1073/pnas.2320489121. Epub 2024 May 28.
7
Synthetic faces generated with the facial action coding system or deep neural networks improve speech-in-noise perception, but not as much as real faces.通过面部动作编码系统或深度神经网络生成的合成面孔可改善噪声环境下的语音感知,但效果不如真实面孔。
Front Neurosci. 2024 May 9;18:1379988. doi: 10.3389/fnins.2024.1379988. eCollection 2024.
8
Hierarchically nested networks optimize the analysis of audiovisual speech.分层嵌套网络优化了视听语音分析。
iScience. 2023 Feb 20;26(3):106257. doi: 10.1016/j.isci.2023.106257. eCollection 2023 Mar 17.
9
The effect of prestimulus low-frequency neural oscillations on the temporal perception of audiovisual speech.刺激前低频神经振荡对视听言语时间感知的影响。
Front Neurosci. 2023 Feb 3;17:1067632. doi: 10.3389/fnins.2023.1067632. eCollection 2023.
10
Differential Auditory and Visual Phase-Locking Are Observed during Audio-Visual Benefit and Silent Lip-Reading for Speech Perception.在听觉-视觉获益和默读唇语感知语音时,观察到听觉和视觉相位锁定的差异。
J Neurosci. 2022 Aug 3;42(31):6108-6120. doi: 10.1523/JNEUROSCI.2476-21.2022. Epub 2022 Jun 27.
刺激前脑振荡相位与跨模态同步知觉相关。
Eur J Neurosci. 2019 Jan;49(2):150-164. doi: 10.1111/ejn.14186. Epub 2018 Oct 24.
4
Joint Representation of Spatial and Phonetic Features in the Human Core Auditory Cortex.人类核心听觉皮层中空间和语音特征的联合表示。
Cell Rep. 2018 Aug 21;24(8):2051-2062.e2. doi: 10.1016/j.celrep.2018.07.076.
5
Representational interactions during audiovisual speech entrainment: Redundancy in left posterior superior temporal gyrus and synergy in left motor cortex.视听语音同步期间的表象相互作用:左后上颞 gyrus 中的冗余和左运动皮层中的协同作用。
PLoS Biol. 2018 Aug 6;16(8):e2006558. doi: 10.1371/journal.pbio.2006558. eCollection 2018 Aug.
6
Electrocorticography reveals continuous auditory and visual speech tracking in temporal and occipital cortex.脑电描记术揭示了颞叶和枕叶皮层中听觉和视觉言语的连续追踪。
Eur J Neurosci. 2020 Mar;51(5):1364-1376. doi: 10.1111/ejn.13992. Epub 2018 Aug 12.
7
The right hemisphere is highlighted in connected natural speech production and perception.右半球在连贯的自然言语产生和感知中被突显出来。
Neuroimage. 2017 May 15;152:628-638. doi: 10.1016/j.neuroimage.2017.03.006. Epub 2017 Mar 6.
8
iELVis: An open source MATLAB toolbox for localizing and visualizing human intracranial electrode data.iELVis:一个用于定位和可视化人类颅内电极数据的开源MATLAB工具箱。
J Neurosci Methods. 2017 Apr 1;281:40-48. doi: 10.1016/j.jneumeth.2017.01.022. Epub 2017 Feb 10.
9
Visual Cortical Entrainment to Motion and Categorical Speech Features during Silent Lipreading.无声唇读过程中视觉皮层对运动和分类语音特征的同步化
Front Hum Neurosci. 2017 Jan 11;10:679. doi: 10.3389/fnhum.2016.00679. eCollection 2016.
10
Perceptual restoration of masked speech in human cortex.人类大脑皮层中掩蔽语音的感知恢复。
Nat Commun. 2016 Dec 20;7:13619. doi: 10.1038/ncomms13619.