• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

独立的时间和语言线索对应机制有益于视听语音处理。

Independent mechanisms of temporal and linguistic cue correspondence benefiting audiovisual speech processing.

机构信息

Brain and Cognitive Sciences, University of Rochester, Rochester, NY, 14627, USA.

Del Monte Institute for Neuroscience, University of Rochester, Rochester, NY, 14627, USA.

出版信息

Atten Percept Psychophys. 2022 Aug;84(6):2016-2026. doi: 10.3758/s13414-022-02440-3. Epub 2022 Feb 24.

DOI:10.3758/s13414-022-02440-3
PMID:35211849
Abstract

It is well established that in order to comprehend speech in noisy environments, listeners use the face of the talker in conjunction with the auditory speech. Yet how listeners use audiovisual speech correspondences along the multisensory speech processing pathway is not known. We engaged listeners in a pair of experiments using face rotation to partially dissociate linguistic and temporal information and two tasks to assess both overall integration and early integration specifically. In our first exploratory experiment, listeners performed a speech in noise task to determine which face rotation maximally disrupts speech comprehension and thus overall audiovisual integration. Our second experiment involved a dual pitch discrimination and visual catch task to test specifically for binding. The results showed that temporal coherence supports early integration, replicating the importance of temporal coherence seen for binding nonspeech stimuli. However, the benefit of temporal coherence was present in both upright and inverted positions, suggesting that binding is minimally affected by face rotation under these conditions. Together, our results suggest that different aspects of audio-visual speech are integrated at different stages of multisensory speech processing.

摘要

众所周知,为了在嘈杂的环境中理解言语,听众会将说话者的面部与听觉言语结合起来使用。然而,听众如何沿着多感官言语处理途径使用视听言语对应关系尚不清楚。我们通过使用面部旋转来部分分离语言和时间信息,让听众参与了两项实验,以评估整体整合和早期整合。在我们的第一个探索性实验中,听众执行了一项噪声中的言语任务,以确定哪种面部旋转最大程度地破坏言语理解,从而整体视听整合。我们的第二个实验涉及双重音高辨别和视觉捕获任务,以专门测试绑定。结果表明,时间连贯性支持早期整合,复制了对于绑定非言语刺激物的时间连贯性重要性。然而,时间连贯性的优势在直立和倒置位置都存在,这表明在这些条件下,面部旋转对绑定的影响最小。总之,我们的结果表明,视听言语的不同方面在多感官言语处理的不同阶段进行整合。

相似文献

1
Independent mechanisms of temporal and linguistic cue correspondence benefiting audiovisual speech processing.独立的时间和语言线索对应机制有益于视听语音处理。
Atten Percept Psychophys. 2022 Aug;84(6):2016-2026. doi: 10.3758/s13414-022-02440-3. Epub 2022 Feb 24.
2
Performance in an Audiovisual Selective Attention Task Using Speech-Like Stimuli Depends on the Talker Identities, But Not Temporal Coherence.使用类似语音的刺激进行视听选择性注意任务的表现取决于说话者的身份,但与时间连贯性无关。
Trends Hear. 2023 Jan-Dec;27:23312165231207235. doi: 10.1177/23312165231207235.
3
Neurophysiological Indices of Audiovisual Speech Processing Reveal a Hierarchy of Multisensory Integration Effects.神经生理指标揭示视听言语加工中的多感觉整合效应层次结构。
J Neurosci. 2021 Jun 9;41(23):4991-5003. doi: 10.1523/JNEUROSCI.0906-20.2021. Epub 2021 Apr 6.
4
Eye Can Hear Clearly Now: Inverse Effectiveness in Natural Audiovisual Speech Processing Relies on Long-Term Crossmodal Temporal Integration.现在眼睛能“听清”了:自然视听言语处理中的反向有效性依赖于长期跨模态时间整合。
J Neurosci. 2016 Sep 21;36(38):9888-95. doi: 10.1523/JNEUROSCI.1396-16.2016.
5
Prediction and constraint in audiovisual speech perception.视听言语感知中的预测与约束
Cortex. 2015 Jul;68:169-81. doi: 10.1016/j.cortex.2015.03.006. Epub 2015 Mar 20.
6
The integration of continuous audio and visual speech in a cocktail-party environment depends on attention.在鸡尾酒会环境中,连续视听语音的整合取决于注意力。
Neuroimage. 2023 Jul 1;274:120143. doi: 10.1016/j.neuroimage.2023.120143. Epub 2023 Apr 29.
7
The timecourse of multisensory speech processing in unilaterally stimulated cochlear implant users revealed by ERPs.通过事件相关电位揭示单侧刺激人工耳蜗使用者多感官语音处理的时程。
Neuroimage Clin. 2022;34:102982. doi: 10.1016/j.nicl.2022.102982. Epub 2022 Mar 4.
8
Physical and perceptual factors shape the neural mechanisms that integrate audiovisual signals in speech comprehension.生理和感知因素塑造了整合言语理解中视听信号的神经机制。
J Neurosci. 2011 Aug 3;31(31):11338-50. doi: 10.1523/JNEUROSCI.6510-10.2011.
9
Deficits in audiovisual speech perception in normal aging emerge at the level of whole-word recognition.正常衰老过程中视听言语感知的缺陷出现在整词识别层面。
Neurobiol Aging. 2015 Jan;36(1):283-91. doi: 10.1016/j.neurobiolaging.2014.08.003. Epub 2014 Aug 7.
10
Generalizable EEG Encoding Models with Naturalistic Audiovisual Stimuli.具有自然视听刺激的可泛化 EEG 编码模型。
J Neurosci. 2021 Oct 27;41(43):8946-8962. doi: 10.1523/JNEUROSCI.2891-20.2021. Epub 2021 Sep 9.

引用本文的文献

1
The effects of temporal cues, point-light displays, and faces on speech identification and listening effort.时线索、点光显示和面部对言语识别和听力努力的影响。
PLoS One. 2023 Nov 29;18(11):e0290826. doi: 10.1371/journal.pone.0290826. eCollection 2023.
2
Performance in an Audiovisual Selective Attention Task Using Speech-Like Stimuli Depends on the Talker Identities, But Not Temporal Coherence.使用类似语音的刺激进行视听选择性注意任务的表现取决于说话者的身份,但与时间连贯性无关。
Trends Hear. 2023 Jan-Dec;27:23312165231207235. doi: 10.1177/23312165231207235.
3
Speech-In-Noise Comprehension is Improved When Viewing a Deep-Neural-Network-Generated Talking Face.

本文引用的文献

1
Visual analog of the acoustic amplitude envelope benefits speech perception in noise.声强度包络的视觉模拟有助于噪声下的言语感知。
J Acoust Soc Am. 2020 Mar;147(3):EL246. doi: 10.1121/10.0000737.
2
Integration of Visual Information in Auditory Cortex Promotes Auditory Scene Analysis through Multisensory Binding.视觉信息在听觉皮层中的整合通过多感觉绑定促进听觉场景分析。
Neuron. 2018 Feb 7;97(3):640-655.e4. doi: 10.1016/j.neuron.2017.12.034. Epub 2018 Jan 26.
3
Individual differences and the effect of face configuration information in the McGurk effect.
观看深度神经网络生成的说话人脸可提高噪声环境下的言语理解能力。
Trends Hear. 2022 Jan-Dec;26:23312165221136934. doi: 10.1177/23312165221136934.
个体差异与麦格克效应中面部构型信息的影响
Exp Brain Res. 2018 Apr;236(4):973-984. doi: 10.1007/s00221-018-5188-4. Epub 2018 Jan 30.
4
The Hierarchical Cortical Organization of Human Speech Processing.人类言语处理的分层皮质组织
J Neurosci. 2017 Jul 5;37(27):6539-6557. doi: 10.1523/JNEUROSCI.3267-16.2017. Epub 2017 Jun 6.
5
Audio-Visual and Meaningful Semantic Context Enhancements in Older and Younger Adults.老年人和年轻人视听及有意义语义语境增强
PLoS One. 2016 Mar 31;11(3):e0152773. doi: 10.1371/journal.pone.0152773. eCollection 2016.
6
Distinct Computational Principles Govern Multisensory Integration in Primary Sensory and Association Cortices.不同的计算原理支配着初级感觉皮层和联合皮层中的多感觉整合。
Curr Biol. 2016 Feb 22;26(4):509-14. doi: 10.1016/j.cub.2015.12.056. Epub 2016 Feb 4.
7
Defining Auditory-Visual Objects: Behavioral Tests and Physiological Mechanisms.定义视听对象:行为测试与生理机制
Trends Neurosci. 2016 Feb;39(2):74-85. doi: 10.1016/j.tins.2015.12.007. Epub 2016 Jan 15.
8
Cortical tracking of hierarchical linguistic structures in connected speech.连贯言语中层次语言结构的皮层追踪。
Nat Neurosci. 2016 Jan;19(1):158-64. doi: 10.1038/nn.4186. Epub 2015 Dec 7.
9
Cortical hierarchies perform Bayesian causal inference in multisensory perception.皮质层级在多感官感知中执行贝叶斯因果推理。
PLoS Biol. 2015 Feb 24;13(2):e1002073. doi: 10.1371/journal.pbio.1002073. eCollection 2015 Feb.
10
Enhancing speech intelligibility: interactions among context, modality, speech style, and masker.提高言语可懂度:语境、模态、言语风格和掩蔽音之间的相互作用。
J Speech Lang Hear Res. 2014 Oct;57(5):1908-18. doi: 10.1044/JSLHR-H-13-0076.