• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

先前的多感官学习可以促进仅听觉模式下的语音身份识别以及噪声环境中的语音识别。

Prior multisensory learning can facilitate auditory-only voice-identity and speech recognition in noise.

作者信息

Maguinness Corrina, Schall Sonja, Mathias Brian, Schoemann Martin, von Kriegstein Katharina

机构信息

Chair of Cognitive and Clinical Neuroscience, Faculty of Psychology, Technische Universität Dresden, Dresden, Germany.

Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany.

出版信息

Q J Exp Psychol (Hove). 2024 Sep 20;78(7):17470218241278649. doi: 10.1177/17470218241278649.

DOI:10.1177/17470218241278649
PMID:39164830
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12181647/
Abstract

Seeing the visual articulatory movements of a speaker, while hearing their voice, helps with understanding what is said. This multisensory enhancement is particularly evident in noisy listening conditions. Multisensory enhancement also occurs even in auditory-only conditions: auditory-only speech and voice-identity recognition are superior for speakers previously learned with their face, compared to control learning; an effect termed the "face-benefit." Whether the face-benefit can assist in maintaining robust perception in increasingly noisy listening conditions, similar to concurrent multisensory input, is unknown. Here, in two behavioural experiments, we examined this hypothesis. In each experiment, participants learned a series of speakers' voices together with their dynamic face or control image. Following learning, participants listened to auditory-only sentences spoken by the same speakers and recognised the content of the sentences (speech recognition, Experiment 1) or the voice-identity of the speaker (Experiment 2) in increasing levels of auditory noise. For speech recognition, we observed that 14 of 30 participants (47%) showed a face-benefit. 19 of 25 participants (76%) showed a face-benefit for voice-identity recognition. For those participants who demonstrated a face-benefit, the face-benefit increased with auditory noise levels. Taken together, the results support an audio-visual model of auditory communication and suggest that the brain can develop a flexible system in which learned facial characteristics are used to deal with varying auditory uncertainty.

摘要

在听说话者声音的同时看到其视觉发音动作,有助于理解所说内容。这种多感官增强在嘈杂的聆听环境中尤为明显。多感官增强甚至在仅听觉条件下也会发生:与对照学习相比,对于之前通过面部学习过的说话者,仅听觉的语音和语音识别能力更强;这种效应被称为“面部优势”。面部优势是否能像同时进行的多感官输入那样,在日益嘈杂的聆听环境中帮助维持强大的感知能力,目前尚不清楚。在此,我们通过两项行为实验对这一假设进行了检验。在每个实验中,参与者将一系列说话者的声音与其动态面部或对照图像一起学习。学习之后,参与者听取由相同说话者说出的仅听觉句子,并在不断增加的听觉噪声水平下识别句子内容(语音识别,实验1)或说话者的语音身份(实验2)。对于语音识别,我们观察到30名参与者中有14名(47%)表现出面部优势。25名参与者中有19名(76%)在语音身份识别方面表现出面部优势。对于那些表现出面部优势的参与者,面部优势随着听觉噪声水平的增加而增强。综上所述,这些结果支持了听觉交流的视听模型,并表明大脑可以发展出一个灵活的系统,其中学习到的面部特征被用于应对不同的听觉不确定性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fe/12181647/fd232e683960/10.1177_17470218241278649-fig5.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fe/12181647/b89d51f725bd/10.1177_17470218241278649-fig1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fe/12181647/6c17b9423232/10.1177_17470218241278649-fig2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fe/12181647/4293362c527e/10.1177_17470218241278649-fig3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fe/12181647/679e6be6f031/10.1177_17470218241278649-fig4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fe/12181647/fd232e683960/10.1177_17470218241278649-fig5.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fe/12181647/b89d51f725bd/10.1177_17470218241278649-fig1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fe/12181647/6c17b9423232/10.1177_17470218241278649-fig2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fe/12181647/4293362c527e/10.1177_17470218241278649-fig3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fe/12181647/679e6be6f031/10.1177_17470218241278649-fig4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/31fe/12181647/fd232e683960/10.1177_17470218241278649-fig5.jpg

相似文献

1
Prior multisensory learning can facilitate auditory-only voice-identity and speech recognition in noise.先前的多感官学习可以促进仅听觉模式下的语音身份识别以及噪声环境中的语音识别。
Q J Exp Psychol (Hove). 2024 Sep 20;78(7):17470218241278649. doi: 10.1177/17470218241278649.
2
Using Pupillometry in Virtual Reality as a Tool for Speech-in-Noise Research.在虚拟现实中使用瞳孔测量法作为噪声环境下语音研究的工具。
Ear Hear. 2025 Jul 2. doi: 10.1097/AUD.0000000000001692.
3
Stigma Management Strategies of Autistic Social Media Users.自闭症社交媒体用户的污名管理策略
Autism Adulthood. 2025 May 28;7(3):273-282. doi: 10.1089/aut.2023.0095. eCollection 2025 Jun.
4
Visual mechanisms for voice-identity recognition flexibly adjust to auditory noise level.视觉语音识别机制可灵活适应听觉噪声水平。
Hum Brain Mapp. 2021 Aug 15;42(12):3963-3982. doi: 10.1002/hbm.25532. Epub 2021 May 27.
5
A systematic review of speech, language and communication interventions for children with Down syndrome from 0 to 6 years.对0至6岁唐氏综合征儿童言语、语言和沟通干预措施的系统评价。
Int J Lang Commun Disord. 2022 Mar;57(2):441-463. doi: 10.1111/1460-6984.12699. Epub 2022 Feb 22.
6
Improving Cognitive Empathy Through Traumatic Brain Injury Experiential Learning: A Novel Mixed Methods Approach for Speech-Language Pathology Graduate Education.通过创伤性脑损伤体验式学习提高认知同理心:一种用于言语语言病理学研究生教育的新型混合方法。
Am J Speech Lang Pathol. 2025 Jan 13:1-33. doi: 10.1044/2024_AJSLP-24-00126.
7
Survivor, family and professional experiences of psychosocial interventions for sexual abuse and violence: a qualitative evidence synthesis.性虐待和暴力的心理社会干预的幸存者、家庭和专业人员的经验:定性证据综合。
Cochrane Database Syst Rev. 2022 Oct 4;10(10):CD013648. doi: 10.1002/14651858.CD013648.pub2.
8
Phonological representations of auditory and visual speech in the occipito-temporal cortex and beyond.枕颞叶皮层及其他区域中听觉和视觉言语的语音表征。
J Neurosci. 2025 Apr 30. doi: 10.1523/JNEUROSCI.1415-24.2025.
9
Improving auditory attention decoding in noisy environments for listeners with hearing impairment through contrastive learning.通过对比学习改善听力受损者在嘈杂环境中的听觉注意力解码。
J Neural Eng. 2025 Jun 18;22(3). doi: 10.1088/1741-2552/ade28a.
10
Effectiveness of voice rehabilitation on vocalisation in postlaryngectomy patients: a systematic review.喉切除术后患者的嗓音康复对发声效果的影响:系统评价。
Int J Evid Based Healthc. 2010 Dec;8(4):256-8. doi: 10.1111/j.1744-1609.2010.00177.x.

本文引用的文献

1
The Benefit of Bimodal Training in Voice Learning.双峰训练在语音学习中的益处。
Brain Sci. 2023 Aug 30;13(9):1260. doi: 10.3390/brainsci13091260.
2
Crossmodal interactions in human learning and memory.人类学习与记忆中的跨模态交互。
Front Hum Neurosci. 2023 May 17;17:1181760. doi: 10.3389/fnhum.2023.1181760. eCollection 2023.
3
Crossmodal plasticity in hearing loss.听力损失的跨模态可塑性。
Trends Neurosci. 2023 May;46(5):377-393. doi: 10.1016/j.tins.2023.02.004. Epub 2023 Mar 27.
4
Enriched learning: behavior, brain, and computation.强化学习:行为、大脑与计算
Trends Cogn Sci. 2023 Jan;27(1):81-97. doi: 10.1016/j.tics.2022.10.007. Epub 2022 Nov 28.
5
Visual mechanisms for voice-identity recognition flexibly adjust to auditory noise level.视觉语音识别机制可灵活适应听觉噪声水平。
Hum Brain Mapp. 2021 Aug 15;42(12):3963-3982. doi: 10.1002/hbm.25532. Epub 2021 May 27.
6
The interrelationship between the face and vocal tract configuration during audiovisual speech.在视听言语中,面部和声道构形之间的相互关系。
Proc Natl Acad Sci U S A. 2020 Dec 22;117(51):32791-32798. doi: 10.1073/pnas.2006192117. Epub 2020 Dec 8.
7
Cross-modal transfer of talker-identity learning.跨模态说话人身份学习的迁移。
Atten Percept Psychophys. 2021 Jan;83(1):415-434. doi: 10.3758/s13414-020-02141-9. Epub 2020 Oct 20.
8
What you see is what you hear: Twenty years of research using the Sound-Induced Flash Illusion.所见即所闻:运用声音诱发闪光错觉的二十年研究
Neurosci Biobehav Rev. 2020 Nov;118:759-774. doi: 10.1016/j.neubiorev.2020.09.006. Epub 2020 Sep 13.
9
Face and Voice Perception: Understanding Commonalities and Differences.面容与声音感知:理解共性与差异。
Trends Cogn Sci. 2020 May;24(5):398-410. doi: 10.1016/j.tics.2020.02.001. Epub 2020 Feb 26.
10
Lip-Reading Enables the Brain to Synthesize Auditory Features of Unknown Silent Speech.唇读使大脑能够合成未知静音语音的听觉特征。
J Neurosci. 2020 Jan 29;40(5):1053-1065. doi: 10.1523/JNEUROSCI.1101-19.2019. Epub 2019 Dec 30.