• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

沉浸式虚拟现实中的同时对话:关于视觉言语线索的主导地位。

Concurrent talking in immersive virtual reality: on the dominance of visual speech cues.

机构信息

Microsoft Research, One Microsoft Way, Redmond, WA, 98052, USA.

Event Lab, Department of Clinical Psychology and Psychobiology, University of Barcelona, Barcelona, 08035, Spain.

出版信息

Sci Rep. 2017 Jun 19;7(1):3817. doi: 10.1038/s41598-017-04201-x.

DOI:10.1038/s41598-017-04201-x
PMID:28630450
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC5476615/
Abstract

Humans are good at selectively listening to specific target conversations, even in the presence of multiple concurrent speakers. In our research, we study how auditory-visual cues modulate this selective listening. We do so by using immersive Virtual Reality technologies with spatialized audio. Exposing 32 participants to an Information Masking Task with concurrent speakers, we find significantly more errors in the decision-making processes triggered by asynchronous audiovisual speech cues. More precisely, the results show that lips on the Target speaker matched to a secondary (Mask) speaker's audio severely increase the participants' comprehension error rates. In a control experiment (n = 20), we further explore the influences of the visual modality over auditory selective attention. The results show a dominance of visual-speech cues, which effectively turn the Mask into the Target and vice-versa. These results reveal a disruption of selective attention that is triggered by bottom-up multisensory integration. The findings are framed in the sensory perception and cognitive neuroscience theories. The VR setup is validated by replicating previous results in this literature in a supplementary experiment.

摘要

人类善于有选择地听取特定的目标对话,即使有多个同时说话的人存在。在我们的研究中,我们通过使用具有空间化音频的沉浸式虚拟现实技术来研究听觉-视觉线索如何调节这种选择性聆听。我们让 32 名参与者参与带有同时说话者的信息掩蔽任务,我们发现,由异步视听语音线索触发的决策过程中的错误明显更多。更准确地说,结果表明,与次要(掩蔽)说话者的音频匹配的目标说话者的嘴唇严重增加了参与者的理解错误率。在一个对照实验(n=20)中,我们进一步探索了视觉模态对听觉选择性注意的影响。结果显示,视觉-语音线索占据主导地位,有效地将掩蔽者变成了目标,反之亦然。这些结果揭示了由自下而上的多感觉整合触发的选择性注意力的破坏。这些发现是在感官知觉和认知神经科学理论的框架内提出的。虚拟现实设置通过在补充实验中复制该文献中的先前结果得到了验证。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e0e2/5476615/7ef4ecfc9ccd/41598_2017_4201_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e0e2/5476615/7e4346d16aee/41598_2017_4201_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e0e2/5476615/64fd3ca340a3/41598_2017_4201_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e0e2/5476615/d8047b619d98/41598_2017_4201_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e0e2/5476615/b0748d75ac1b/41598_2017_4201_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e0e2/5476615/7ef4ecfc9ccd/41598_2017_4201_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e0e2/5476615/7e4346d16aee/41598_2017_4201_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e0e2/5476615/64fd3ca340a3/41598_2017_4201_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e0e2/5476615/d8047b619d98/41598_2017_4201_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e0e2/5476615/b0748d75ac1b/41598_2017_4201_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e0e2/5476615/7ef4ecfc9ccd/41598_2017_4201_Fig5_HTML.jpg

相似文献

1
Concurrent talking in immersive virtual reality: on the dominance of visual speech cues.沉浸式虚拟现实中的同时对话:关于视觉言语线索的主导地位。
Sci Rep. 2017 Jun 19;7(1):3817. doi: 10.1038/s41598-017-04201-x.
2
The Dynamics of Attention Shifts Among Concurrent Speech in a Naturalistic Multi-speaker Virtual Environment.自然主义多说话者虚拟环境中并行语音间注意力转移的动态变化
Front Hum Neurosci. 2019 Nov 8;13:386. doi: 10.3389/fnhum.2019.00386. eCollection 2019.
3
Intentional switching in auditory selective attention: Exploring age-related effects in a spatial setup requiring speech perception.听觉选择性注意中的有意切换:在需要语音感知的空间设置中探索年龄相关效应。
Acta Psychol (Amst). 2017 Jun;177:36-43. doi: 10.1016/j.actpsy.2017.04.008. Epub 2017 Apr 28.
4
Perception of co-speech gestures in aphasic patients: a visual exploration study during the observation of dyadic conversations.失语症患者对伴随言语手势的感知:二元对话观察期间的视觉探索性研究
Cortex. 2015 Mar;64:157-68. doi: 10.1016/j.cortex.2014.10.013. Epub 2014 Nov 4.
5
Prediction and constraint in audiovisual speech perception.视听言语感知中的预测与约束
Cortex. 2015 Jul;68:169-81. doi: 10.1016/j.cortex.2015.03.006. Epub 2015 Mar 20.
6
Reaching to sounds in virtual reality: A multisensory-motor approach to promote adaptation to altered auditory cues.虚拟现实中的声音感知:促进适应改变的听觉线索的多感觉运动方法。
Neuropsychologia. 2020 Dec;149:107665. doi: 10.1016/j.neuropsychologia.2020.107665. Epub 2020 Oct 29.
7
Top-down attention regulates the neural expression of audiovisual integration.自上而下的注意力调节视听整合的神经表达。
Neuroimage. 2015 Oct 1;119:272-85. doi: 10.1016/j.neuroimage.2015.06.052. Epub 2015 Jun 26.
8
The effects of substitute multisensory feedback on task performance and the sense of presence in a virtual reality environment.替代性多感官反馈对虚拟现实环境中任务表现和临场感的影响。
PLoS One. 2018 Feb 1;13(2):e0191846. doi: 10.1371/journal.pone.0191846. eCollection 2018.
9
Older adults expend more listening effort than young adults recognizing audiovisual speech in noise.老年人在噪声中识别视听语音比年轻人需要更多的听力努力。
Int J Audiol. 2011 Nov;50(11):786-92. doi: 10.3109/14992027.2011.599870. Epub 2011 Sep 15.
10
When eyes beat lips: speaker gaze affects audiovisual integration in the McGurk illusion.当眼睛打败嘴唇:说话者注视影响 McGurk 错觉中的视听整合。
Psychol Res. 2022 Sep;86(6):1930-1943. doi: 10.1007/s00426-021-01618-y. Epub 2021 Dec 2.

引用本文的文献

1
Speaker-story mapping as a method to evaluate audiovisual scene analysis in a virtual classroom scenario.将讲述者-故事映射作为一种在虚拟课堂场景中评估视听场景分析的方法。
Front Psychol. 2025 Jun 10;16:1520630. doi: 10.3389/fpsyg.2025.1520630. eCollection 2025.
2
Immersive interfaces for clinical applications: current status and future perspective.临床应用中的沉浸式界面:现状与未来展望。
Front Neurorobot. 2024 Nov 27;18:1362444. doi: 10.3389/fnbot.2024.1362444. eCollection 2024.
3
Speech Intelligibility in Reverberation is Reduced During Self-Rotation.

本文引用的文献

1
The sense of body ownership relaxes temporal constraints for multisensory integration.身体所有权感放宽了多感官整合的时间限制。
Sci Rep. 2016 Aug 3;6:30628. doi: 10.1038/srep30628.
2
Cross-modal Informational Masking of Lipreading by Babble.嘈杂语音对唇读的跨模态信息掩蔽
Atten Percept Psychophys. 2016 Jan;78(1):346-54. doi: 10.3758/s13414-015-0990-6.
3
Violating body movement semantics: Neural signatures of self-generated and external-generated errors.违反身体运动语义:自我产生和外部产生错误的神经特征。
自旋转时混响中的语音可懂度降低。
Trends Hear. 2023 Jan-Dec;27:23312165231188619. doi: 10.1177/23312165231188619.
4
Feasibility of Virtual Reality Audiological Testing: Prospective Study.虚拟现实听力测试的可行性:前瞻性研究。
JMIR Serious Games. 2021 Aug 31;9(3):e26976. doi: 10.2196/26976.
5
Rethinking GPS navigation: creating cognitive maps through auditory clues.重新思考 GPS 导航:通过听觉线索创建认知地图。
Sci Rep. 2021 Apr 8;11(1):7764. doi: 10.1038/s41598-021-87148-4.
6
Sensitivity of occipito-temporal cortex, premotor and Broca's areas to visible speech gestures in a familiar language.对熟悉语言的可视言语手势,枕颞皮质、运动前区和布罗卡区的敏感性。
PLoS One. 2020 Jun 19;15(6):e0234695. doi: 10.1371/journal.pone.0234695. eCollection 2020.
7
Temporal contrast effects in human speech perception are immune to selective attention.人类言语感知中的时间对比效应不受选择性注意的影响。
Sci Rep. 2020 Mar 27;10(1):5607. doi: 10.1038/s41598-020-62613-8.
8
The Influence of Auditory Cues on Bodily and Movement Perception.听觉线索对身体和运动感知的影响。
Front Psychol. 2020 Jan 17;10:3001. doi: 10.3389/fpsyg.2019.03001. eCollection 2019.
9
Generic HRTFs May be Good Enough in Virtual Reality. Improving Source Localization through Cross-Modal Plasticity.通用头部相关传递函数在虚拟现实中可能就足够好了。通过跨模态可塑性改善声源定位。
Front Neurosci. 2018 Feb 2;12:21. doi: 10.3389/fnins.2018.00021. eCollection 2018.
Neuroimage. 2016 Jan 1;124(Pt A):147-156. doi: 10.1016/j.neuroimage.2015.08.022. Epub 2015 Aug 15.
4
Selective spatial attention modulates bottom-up informational masking of speech.选择性空间注意调节语音的自下而上信息掩蔽。
Sci Rep. 2015 Mar 2;5:8662. doi: 10.1038/srep08662.
5
Body ownership causes illusory self-attribution of speaking and influences subsequent real speaking.身体所有权会导致对言语的虚幻自我归因,并影响随后的实际言语。
Proc Natl Acad Sci U S A. 2014 Dec 9;111(49):17678-83. doi: 10.1073/pnas.1414936111. Epub 2014 Nov 24.
6
A threat to a virtual hand elicits motor cortex activation.对虚拟手的威胁会引发运动皮层激活。
Exp Brain Res. 2014 Mar;232(3):875-87. doi: 10.1007/s00221-013-3800-1. Epub 2013 Dec 13.
7
Linguistic contributions to speech-on-speech masking for native and non-native listeners: language familiarity and semantic content.母语和非母语听者的言语掩蔽中的语言贡献:语言熟悉度和语义内容。
J Acoust Soc Am. 2012 Feb;131(2):1449-64. doi: 10.1121/1.3675943.
8
Crossmodal influences on visual perception.跨模态对视觉感知的影响。
Phys Life Rev. 2010 Sep;7(3):269-84. doi: 10.1016/j.plrev.2010.04.006. Epub 2010 Apr 22.
9
The cocktail party problem.鸡尾酒会问题。
Curr Biol. 2009 Dec 1;19(22):R1024-7. doi: 10.1016/j.cub.2009.09.005.
10
Place illusion and plausibility can lead to realistic behaviour in immersive virtual environments.沉浸虚拟环境中的-place 错觉和合理性可以导致现实行为。
Philos Trans R Soc Lond B Biol Sci. 2009 Dec 12;364(1535):3549-57. doi: 10.1098/rstb.2009.0138.