• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

言语感知中的自动视听整合。

Automatic audiovisual integration in speech perception.

作者信息

Gentilucci Maurizio, Cattaneo Luigi

机构信息

Dipartimento di Neuroscienze, Universitá di Parma, Via Volturno 39, 43100, Parma, Italy.

出版信息

Exp Brain Res. 2005 Nov;167(1):66-75. doi: 10.1007/s00221-005-0008-z. Epub 2005 Oct 29.

DOI:10.1007/s00221-005-0008-z
PMID:16034571
Abstract

Two experiments aimed to determine whether features of both the visual and acoustical inputs are always merged into the perceived representation of speech and whether this audiovisual integration is based on either cross-modal binding functions or on imitation. In a McGurk paradigm, observers were required to repeat aloud a string of phonemes uttered by an actor (acoustical presentation of phonemic string) whose mouth, in contrast, mimicked pronunciation of a different string (visual presentation). In a control experiment participants read the same printed strings of letters. This condition aimed to analyze the pattern of voice and the lip kinematics controlling for imitation. In the control experiment and in the congruent audiovisual presentation, i.e. when the articulation mouth gestures were congruent with the emission of the string of phones, the voice spectrum and the lip kinematics varied according to the pronounced strings of phonemes. In the McGurk paradigm the participants were unaware of the incongruence between visual and acoustical stimuli. The acoustical analysis of the participants' spoken responses showed three distinct patterns: the fusion of the two stimuli (the McGurk effect), repetition of the acoustically presented string of phonemes, and, less frequently, of the string of phonemes corresponding to the mouth gestures mimicked by the actor. However, the analysis of the latter two responses showed that the formant 2 of the participants' voice spectra always differed from the value recorded in the congruent audiovisual presentation. It approached the value of the formant 2 of the string of phonemes presented in the other modality, which was apparently ignored. The lip kinematics of the participants repeating the string of phonemes acoustically presented were influenced by the observation of the lip movements mimicked by the actor, but only when pronouncing a labial consonant. The data are discussed in favor of the hypothesis that features of both the visual and acoustical inputs always contribute to the representation of a string of phonemes and that cross-modal integration occurs by extracting mouth articulation features peculiar for the pronunciation of that string of phonemes.

摘要

两项实验旨在确定视觉和听觉输入的特征是否总是融合到语音的感知表征中,以及这种视听整合是基于跨模态绑定功能还是基于模仿。在麦格克范式中,要求观察者大声重复演员说出的一串音素(音素串的听觉呈现),而演员的嘴则模仿不同音素串的发音(视觉呈现)。在一个对照实验中,参与者阅读相同的印刷字母串。该条件旨在分析语音模式和控制模仿的唇运动学。在对照实验和一致的视听呈现中,即当发音的口部手势与音素串的发出一致时,语音频谱和唇运动学根据发音的音素串而变化。在麦格克范式中,参与者没有意识到视觉和听觉刺激之间的不一致。对参与者口语反应的声学分析显示出三种不同的模式:两种刺激的融合(麦格克效应)、重复听觉呈现的音素串,以及较少出现的与演员模仿的口部手势对应的音素串。然而,对后两种反应的分析表明,参与者语音频谱的第二共振峰总是与在一致的视听呈现中记录的值不同。它接近在另一种模态中呈现的音素串的第二共振峰的值,而这个值显然被忽略了。重复听觉呈现的音素串的参与者的唇运动学受到对演员模仿的唇运动观察的影响,但仅在发唇音辅音时。讨论这些数据支持这样的假设,即视觉和听觉输入的特征总是有助于音素串的表征,并且跨模态整合是通过提取该音素串发音特有的口部发音特征来实现的。

相似文献

1
Automatic audiovisual integration in speech perception.言语感知中的自动视听整合。
Exp Brain Res. 2005 Nov;167(1):66-75. doi: 10.1007/s00221-005-0008-z. Epub 2005 Oct 29.
2
Imitation during phoneme production.音素产生过程中的模仿。
Neuropsychologia. 2007 Feb 1;45(3):608-15. doi: 10.1016/j.neuropsychologia.2006.04.004. Epub 2006 May 12.
3
Degradation of labial information modifies audiovisual speech perception in cochlear-implanted children.唇语信息的退化改变了植入人工耳蜗的儿童对视听语音的感知。
Ear Hear. 2013 Jan-Feb;34(1):110-21. doi: 10.1097/AUD.0b013e3182670993.
4
Audiovisual perception of congruent and incongruent Dutch front vowels.荷兰语前元音的和谐与不和谐的视听感知。
J Speech Lang Hear Res. 2012 Dec;55(6):1788-801. doi: 10.1044/1092-4388(2012/11-0227). Epub 2012 Sep 19.
5
Electrophysiological evidence for speech-specific audiovisual integration.言语特异性视听整合的电生理学证据。
Neuropsychologia. 2014 Jan;53:115-21. doi: 10.1016/j.neuropsychologia.2013.11.011. Epub 2013 Nov 27.
6
Attention to touch weakens audiovisual speech integration.对触觉的关注会削弱视听言语整合。
Exp Brain Res. 2007 Nov;183(3):399-404. doi: 10.1007/s00221-007-1110-1.
7
Audiovisual speech perception in children with developmental language disorder in degraded listening conditions.在听力下降的情况下,发育性语言障碍儿童的视听言语感知。
J Speech Lang Hear Res. 2013 Feb;56(1):211-21. doi: 10.1044/1092-4388(2012/11-0270). Epub 2012 May 31.
8
Prediction across sensory modalities: A neurocomputational model of the McGurk effect.跨感觉通道的预测:麦格克效应的神经计算模型。
Cortex. 2015 Jul;68:61-75. doi: 10.1016/j.cortex.2015.04.008. Epub 2015 Apr 30.
9
"Paying" attention to audiovisual speech: Do incongruent stimuli incur greater costs?“关注”视听言语:不一致的刺激会带来更高的代价吗?
Atten Percept Psychophys. 2019 Aug;81(6):1743-1756. doi: 10.3758/s13414-019-01772-x.
10
Perceptual uncertainty explains activation differences between audiovisual congruent speech and McGurk stimuli.感知不确定性解释了视听一致的语音和麦格克刺激之间激活差异的原因。
Hum Brain Mapp. 2024 Mar;45(4):e26653. doi: 10.1002/hbm.26653.

引用本文的文献

1
Audiovisual integration of rhythm in musicians and dancers.音乐人和舞者的节奏视听整合。
Atten Percept Psychophys. 2024 May;86(4):1400-1416. doi: 10.3758/s13414-024-02874-x. Epub 2024 Apr 1.
2
Multisensory Integration as per Technological Advances: A Review.基于技术进步的多感官整合:综述
Front Neurosci. 2021 Jun 22;15:652611. doi: 10.3389/fnins.2021.652611. eCollection 2021.
3
Animated virtual characters to explore audio-visual speech in controlled and naturalistic environments.用于在受控和自然环境中探索视听语音的动画虚拟角色。

本文引用的文献

1
Mandarin speech perception by ear and eye follows a universal principle.通过耳朵和眼睛进行的普通话语音感知遵循一个普遍原则。
Percept Psychophys. 2004 Jul;66(5):820-36. doi: 10.3758/bf03194976.
2
Modulation of motor excitability during speech perception: the role of Broca's area.言语感知过程中运动兴奋性的调节:布洛卡区的作用。
J Cogn Neurosci. 2004 Jul-Aug;16(6):978-87. doi: 10.1162/0898929041502616.
3
Neural circuits involved in the recognition of actions performed by nonconspecifics: an FMRI study.参与识别非同种个体所执行动作的神经回路:一项功能磁共振成像研究。
Sci Rep. 2020 Sep 23;10(1):15540. doi: 10.1038/s41598-020-72375-y.
4
A Laboratory Study of the McGurk Effect in 324 Monozygotic and Dizygotic Twins.对324对同卵和异卵双胞胎的麦格克效应的实验室研究。
Front Neurosci. 2019 Oct 4;13:1029. doi: 10.3389/fnins.2019.01029. eCollection 2019.
5
What accounts for individual differences in susceptibility to the McGurk effect?个体对麦格克效应的易感性差异的原因是什么?
PLoS One. 2018 Nov 12;13(11):e0207160. doi: 10.1371/journal.pone.0207160. eCollection 2018.
6
Skilled musicians are not subject to the McGurk effect.技艺娴熟的音乐家不会受到麦格克效应的影响。
Sci Rep. 2016 Jul 25;6:30423. doi: 10.1038/srep30423.
7
From Mimicry to Language: A Neuroanatomically Based Evolutionary Model of the Emergence of Vocal Language.从模仿到语言:基于神经解剖学的有声语言出现的进化模型
Front Neurosci. 2016 Jun 30;10:307. doi: 10.3389/fnins.2016.00307. eCollection 2016.
8
Please say what this word is-Vowel-extrinsic normalization in the sensorimotor control of speech.请说出这个词——言语感觉运动控制中的元音外在归一化。
J Exp Psychol Hum Percept Perform. 2016 Jul;42(7):1039-47. doi: 10.1037/xhp0000209. Epub 2016 Jan 28.
9
The Bilingual Language Interaction Network for Comprehension of Speech.用于言语理解的双语语言交互网络
Biling (Camb Engl). 2013 Apr 1;16(2). doi: 10.1017/S1366728912000466.
10
Speech Perception as a Multimodal Phenomenon.作为一种多模态现象的言语感知
Curr Dir Psychol Sci. 2008 Dec;17(6):405-409. doi: 10.1111/j.1467-8721.2008.00615.x.
J Cogn Neurosci. 2004 Jan-Feb;16(1):114-26. doi: 10.1162/089892904322755601.
4
Functional imaging of face and hand imitation: towards a motor theory of empathy.面部与手部模仿的功能成像:迈向共情的运动理论
Neuroimage. 2004 Feb;21(2):601-7. doi: 10.1016/j.neuroimage.2003.09.038.
5
Execution and observation of bringing a fruit to the mouth affect syllable pronunciation.将水果送到嘴边的动作执行与观察会影响音节发音。
Eur J Neurosci. 2004 Jan;19(1):190-202. doi: 10.1111/j.1460-9568.2004.03104.x.
6
Auditory-visual speech perception examined by fMRI and PET.通过功能磁共振成像(fMRI)和正电子发射断层扫描(PET)检查视听语音感知。
Neurosci Res. 2003 Nov;47(3):277-87. doi: 10.1016/s0168-0102(03)00214-1.
7
Activations related to "mirror" and "canonical" neurones in the human brain: an fMRI study.人类大脑中与“镜像”和“典型”神经元相关的激活:一项功能磁共振成像研究。
Neuroimage. 2003 Apr;18(4):928-37. doi: 10.1016/s1053-8119(03)00042-9.
8
Neural mechanisms of empathy in humans: a relay from neural systems for imitation to limbic areas.人类共情的神经机制:从模仿神经系统到边缘区域的传递
Proc Natl Acad Sci U S A. 2003 Apr 29;100(9):5497-502. doi: 10.1073/pnas.0935845100. Epub 2003 Apr 7.
9
The essential role of Broca's area in imitation.布洛卡区在模仿中的重要作用。
Eur J Neurosci. 2003 Mar;17(5):1123-8. doi: 10.1046/j.1460-9568.2003.02530.x.
10
Reading speech from still and moving faces: the neural substrates of visible speech.从静止和动态面部读取语音:可见语音的神经基础。
J Cogn Neurosci. 2003 Jan 1;15(1):57-70. doi: 10.1162/089892903321107828.