• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

在鸡尾酒会场景中,注视对多感官整合的脑电图测量的影响。

The effect of gaze on EEG measures of multisensory integration in a cocktail party scenario.

作者信息

Ahmed Farhin, Nidiffer Aaron R, Lalor Edmund C

机构信息

Department of Biomedical Engineering, Department of Neuroscience, and Del Monte Institute for Neuroscience, and Center for Visual Science, University of Rochester, Rochester, NY, United States.

出版信息

Front Hum Neurosci. 2023 Dec 15;17:1283206. doi: 10.3389/fnhum.2023.1283206. eCollection 2023.

DOI:10.3389/fnhum.2023.1283206
PMID:38162285
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10754997/
Abstract

Seeing the speaker's face greatly improves our speech comprehension in noisy environments. This is due to the brain's ability to combine the auditory and the visual information around us, a process known as multisensory integration. Selective attention also strongly influences what we comprehend in scenarios with multiple speakers-an effect known as the cocktail-party phenomenon. However, the interaction between attention and multisensory integration is not fully understood, especially when it comes to natural, continuous speech. In a recent electroencephalography (EEG) study, we explored this issue and showed that multisensory integration is enhanced when an audiovisual speaker is attended compared to when that speaker is unattended. Here, we extend that work to investigate how this interaction varies depending on a person's gaze behavior, which affects the quality of the visual information they have access to. To do so, we recorded EEG from 31 healthy adults as they performed selective attention tasks in several paradigms involving two concurrently presented audiovisual speakers. We then modeled how the recorded EEG related to the audio speech (envelope) of the presented speakers. Crucially, we compared two classes of model - one that assumed underlying multisensory integration (AV) versus another that assumed two independent unisensory audio and visual processes (A+V). This comparison revealed evidence of strong attentional effects on multisensory integration when participants were looking directly at the face of an audiovisual speaker. This effect was not apparent when the speaker's face was in the peripheral vision of the participants. Overall, our findings suggest a strong influence of attention on multisensory integration when high fidelity visual (articulatory) speech information is available. More generally, this suggests that the interplay between attention and multisensory integration during natural audiovisual speech is dynamic and is adaptable based on the specific task and environment.

摘要

在嘈杂环境中,看到说话者的脸能极大地提高我们对言语的理解。这是因为大脑有能力将我们周围的听觉和视觉信息结合起来,这一过程被称为多感官整合。选择性注意也会强烈影响我们在有多个说话者的场景中的理解,这种效应被称为鸡尾酒会现象。然而,注意力和多感官整合之间的相互作用尚未完全被理解,尤其是在涉及自然、连续言语的情况下。在最近的一项脑电图(EEG)研究中,我们探讨了这个问题,并表明与未被关注的视听说话者相比,当一个视听说话者被关注时,多感官整合会增强。在这里,我们扩展这项工作,以研究这种相互作用如何根据一个人的注视行为而变化,注视行为会影响他们所能获得的视觉信息的质量。为此,我们记录了31名健康成年人在涉及两个同时呈现的视听说话者的几个范式中执行选择性注意任务时的脑电图。然后,我们对记录的脑电图与所呈现说话者的音频语音(包络)之间的关系进行了建模。至关重要的是,我们比较了两类模型——一类假设存在潜在的多感官整合(AV),另一类假设存在两个独立的单感官音频和视觉过程(A+V)。这种比较揭示了参与者直接看着视听说话者的脸时,注意力对多感官整合有强烈影响的证据。当说话者的脸处于参与者的周边视野时,这种效应并不明显。总体而言,我们的研究结果表明,当有高保真视觉(发音)语音信息时,注意力对多感官整合有强烈影响。更普遍地说,这表明在自然视听言语过程中,注意力和多感官整合之间的相互作用是动态的,并且可以根据特定任务和环境进行调整。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8f28/10754997/91b6f7cbfe6b/fnhum-17-1283206-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8f28/10754997/21954c0fe146/fnhum-17-1283206-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8f28/10754997/aae295335b78/fnhum-17-1283206-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8f28/10754997/f128cae59529/fnhum-17-1283206-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8f28/10754997/469f907a85db/fnhum-17-1283206-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8f28/10754997/91b6f7cbfe6b/fnhum-17-1283206-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8f28/10754997/21954c0fe146/fnhum-17-1283206-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8f28/10754997/aae295335b78/fnhum-17-1283206-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8f28/10754997/f128cae59529/fnhum-17-1283206-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8f28/10754997/469f907a85db/fnhum-17-1283206-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8f28/10754997/91b6f7cbfe6b/fnhum-17-1283206-g005.jpg

相似文献

1
The effect of gaze on EEG measures of multisensory integration in a cocktail party scenario.在鸡尾酒会场景中,注视对多感官整合的脑电图测量的影响。
Front Hum Neurosci. 2023 Dec 15;17:1283206. doi: 10.3389/fnhum.2023.1283206. eCollection 2023.
2
The effect of gaze on EEG measures of multisensory integration in a cocktail party scenario.在鸡尾酒会场景中,注视对多感官整合的脑电图测量的影响。
bioRxiv. 2023 Aug 24:2023.08.23.554451. doi: 10.1101/2023.08.23.554451.
3
The integration of continuous audio and visual speech in a cocktail-party environment depends on attention.在鸡尾酒会环境中,连续视听语音的整合取决于注意力。
Neuroimage. 2023 Jul 1;274:120143. doi: 10.1016/j.neuroimage.2023.120143. Epub 2023 Apr 29.
4
Look at me when I'm talking to you: Selective attention at a multisensory cocktail party can be decoded using stimulus reconstruction and alpha power modulations.当我和你说话时看着我:使用刺激重建和阿尔法功率调制,可以对多感官鸡尾酒会上的选择性注意进行解码。
Eur J Neurosci. 2019 Oct;50(8):3282-3295. doi: 10.1111/ejn.14425. Epub 2019 May 17.
5
Eye Can Hear Clearly Now: Inverse Effectiveness in Natural Audiovisual Speech Processing Relies on Long-Term Crossmodal Temporal Integration.现在眼睛能“听清”了:自然视听言语处理中的反向有效性依赖于长期跨模态时间整合。
J Neurosci. 2016 Sep 21;36(38):9888-95. doi: 10.1523/JNEUROSCI.1396-16.2016.
6
Neurophysiological Indices of Audiovisual Speech Processing Reveal a Hierarchy of Multisensory Integration Effects.神经生理指标揭示视听言语加工中的多感觉整合效应层次结构。
J Neurosci. 2021 Jun 9;41(23):4991-5003. doi: 10.1523/JNEUROSCI.0906-20.2021. Epub 2021 Apr 6.
7
Congruent Visual Speech Enhances Cortical Entrainment to Continuous Auditory Speech in Noise-Free Conditions.在无噪声条件下,匹配的视觉语音增强了皮质对连续听觉语音的同步化。
J Neurosci. 2015 Oct 21;35(42):14195-204. doi: 10.1523/JNEUROSCI.1829-15.2015.
8
Left Motor δ Oscillations Reflect Asynchrony Detection in Multisensory Speech Perception.左运动 δ 振荡反映多感觉语音感知中的异步检测。
J Neurosci. 2022 Mar 16;42(11):2313-2326. doi: 10.1523/JNEUROSCI.2965-20.2022. Epub 2022 Jan 27.
9
When eyes beat lips: speaker gaze affects audiovisual integration in the McGurk illusion.当眼睛打败嘴唇:说话者注视影响 McGurk 错觉中的视听整合。
Psychol Res. 2022 Sep;86(6):1930-1943. doi: 10.1007/s00426-021-01618-y. Epub 2021 Dec 2.
10
Top-down attention regulates the neural expression of audiovisual integration.自上而下的注意力调节视听整合的神经表达。
Neuroimage. 2015 Oct 1;119:272-85. doi: 10.1016/j.neuroimage.2015.06.052. Epub 2015 Jun 26.

引用本文的文献

1
Neural Speech Tracking during Selective Attention: A Spatially Realistic Audiovisual Study.选择性注意期间的神经语音追踪:一项空间逼真的视听研究。
eNeuro. 2025 Jun 24;12(6). doi: 10.1523/ENEURO.0132-24.2025. Print 2025 Jun.
2
Effects of multisensory simultaneity judgment training on the comprehension and cortical processing of speech in noise: a randomized controlled trial.多感官同步判断训练对噪声环境中语音理解及皮层加工的影响:一项随机对照试验
Sci Rep. 2025 Apr 15;15(1):12956. doi: 10.1038/s41598-025-96121-4.

本文引用的文献

1
Auditory cortex encodes lipreading information through spatially distributed activity.听觉皮层通过空间分布的活动来编码唇读信息。
Curr Biol. 2024 Sep 9;34(17):4021-4032.e5. doi: 10.1016/j.cub.2024.07.073. Epub 2024 Aug 16.
2
The integration of continuous audio and visual speech in a cocktail-party environment depends on attention.在鸡尾酒会环境中,连续视听语音的整合取决于注意力。
Neuroimage. 2023 Jul 1;274:120143. doi: 10.1016/j.neuroimage.2023.120143. Epub 2023 Apr 29.
3
Interaction of bottom-up and top-down neural mechanisms in spatial multi-talker speech perception.
上下行神经机制在空间多说话人语音感知中的相互作用。
Curr Biol. 2022 Sep 26;32(18):3971-3986.e4. doi: 10.1016/j.cub.2022.07.047. Epub 2022 Aug 15.
4
Look at me when I'm talking to you: Selective attention at a multisensory cocktail party can be decoded using stimulus reconstruction and alpha power modulations.当我和你说话时看着我:使用刺激重建和阿尔法功率调制,可以对多感官鸡尾酒会上的选择性注意进行解码。
Eur J Neurosci. 2019 Oct;50(8):3282-3295. doi: 10.1111/ejn.14425. Epub 2019 May 17.
5
Specific lexico-semantic predictions are associated with unique spatial and temporal patterns of neural activity.特定的词汇语义预测与独特的空间和时间模式的神经活动有关。
Elife. 2018 Dec 21;7:e39061. doi: 10.7554/eLife.39061.
6
Multisensory perception reflects individual differences in processing temporal correlations.多感觉感知反映了处理时间相关性的个体差异。
Sci Rep. 2018 Sep 27;8(1):14483. doi: 10.1038/s41598-018-32673-y.
7
Electrophysiological Correlates of Semantic Dissimilarity Reflect the Comprehension of Natural, Narrative Speech.语义相似度的电生理相关性反映了对自然、叙事性言语的理解。
Curr Biol. 2018 Mar 5;28(5):803-809.e3. doi: 10.1016/j.cub.2018.01.080. Epub 2018 Feb 22.
8
Integration of Visual Information in Auditory Cortex Promotes Auditory Scene Analysis through Multisensory Binding.视觉信息在听觉皮层中的整合通过多感觉绑定促进听觉场景分析。
Neuron. 2018 Feb 7;97(3):640-655.e4. doi: 10.1016/j.neuron.2017.12.034. Epub 2018 Jan 26.
9
Cortical Representations of Speech in a Multitalker Auditory Scene.多说话者听觉场景中语音的皮质表征
J Neurosci. 2017 Sep 20;37(38):9189-9196. doi: 10.1523/JNEUROSCI.0938-17.2017. Epub 2017 Aug 18.
10
The Multivariate Temporal Response Function (mTRF) Toolbox: A MATLAB Toolbox for Relating Neural Signals to Continuous Stimuli.多元时间响应函数(mTRF)工具箱:一个用于将神经信号与连续刺激相关联的MATLAB工具箱。
Front Hum Neurosci. 2016 Nov 30;10:604. doi: 10.3389/fnhum.2016.00604. eCollection 2016.