Suppr超能文献

跨模态语音编码的神经机制。

Neural Mechanisms Underlying Cross-Modal Phonetic Encoding.

机构信息

Center for Mind and Brain, University of California, Davis, California 95618, and

Center for Mind and Brain, University of California, Davis, California 95618, and.

出版信息

J Neurosci. 2018 Feb 14;38(7):1835-1849. doi: 10.1523/JNEUROSCI.1566-17.2017. Epub 2017 Dec 20.

Abstract

Audiovisual (AV) integration is essential for speech comprehension, especially in adverse listening situations. Divergent, but not mutually exclusive, theories have been proposed to explain the neural mechanisms underlying AV integration. One theory advocates that this process occurs via interactions between the auditory and visual cortices, as opposed to fusion of AV percepts in a multisensory integrator. Building upon this idea, we proposed that AV integration in spoken language reflects visually induced weighting of phonetic representations at the auditory cortex. EEG was recorded while male and female human subjects watched and listened to videos of a speaker uttering consonant vowel (CV) syllables /ba/ and /fa/, presented in Auditory-only, AV congruent or incongruent contexts. Subjects reported whether they heard /ba/ or /fa/. We hypothesized that vision alters phonetic encoding by dynamically weighting which phonetic representation in the auditory cortex is strengthened or weakened. That is, when subjects are presented with visual /fa/ and acoustic /ba/ and hear /fa/ (), the visual input strengthens the weighting of the phone /f/ representation. When subjects are presented with visual /ba/ and acoustic /fa/ and hear /ba/ (), the visual input weakens the weighting of the phone /f/ representation. Indeed, we found an enlarged N1 auditory evoked potential when subjects perceived , and a reduced N1 when they perceived , mirroring the N1 behavior for /ba/ and /fa/ in Auditory-only settings. These effects were especially pronounced in individuals with more robust illusory perception. These findings provide evidence that visual speech modifies phonetic encoding at the auditory cortex. The current study presents evidence that audiovisual integration in spoken language occurs when one modality (vision) acts on representations of a second modality (audition). Using the McGurk illusion, we show that visual context primes phonetic representations at the auditory cortex, altering the auditory percept, evidenced by changes in the N1 auditory evoked potential. This finding reinforces the theory that audiovisual integration occurs via visual networks influencing phonetic representations in the auditory cortex. We believe that this will lead to the generation of new hypotheses regarding cross-modal mapping, particularly whether it occurs via direct or indirect routes (e.g., via a multisensory mediator).

摘要

视听(AV)整合对于言语理解至关重要,尤其是在不利的听力环境中。虽然存在一些不同的理论,但它们都试图解释视听整合的神经机制。其中一个理论认为,这个过程是通过听觉和视觉皮层之间的相互作用来实现的,而不是通过在多感觉整合器中融合视听感知来实现的。在此基础上,我们提出,口语中的 AV 整合反映了视觉对听觉皮层中语音表征的加权作用。当男性和女性被试观看并聆听说话者说出辅音-元音(CV)音节 /ba/ 和 /fa/ 的视频时,我们记录了他们的脑电图(EEG)。这些视频分别在听觉呈现、视听一致和不一致的情况下呈现。被试报告他们听到的是 /ba/ 还是 /fa/。我们假设,视觉通过动态地加权听觉皮层中哪个语音表征得到加强或减弱来改变语音编码。也就是说,当被试看到视觉上的 /fa/ 和听觉上的 /ba/ 并听到 /fa/()时,视觉输入会加强对 /f/ 音的表示的权重。当被试看到视觉上的 /ba/ 和听觉上的 /fa/ 并听到 /ba/()时,视觉输入会削弱对 /f/ 音的表示的权重。事实上,当被试感知到 /fa/ 时,我们发现听觉诱发 N1 电位增大,而当被试感知到 /ba/ 时,N1 电位减小,这与听觉环境中 /ba/ 和 /fa/ 的 N1 行为相匹配。在具有更强幻觉感知的个体中,这些效应更为明显。这些发现提供了证据表明,视觉语音会改变听觉皮层中的语音编码。本研究提供的证据表明,当一种模态(视觉)作用于另一种模态(听觉)的表示时,口语中的视听整合就会发生。通过使用麦格克错觉,我们表明,视觉上下文可以在听觉皮层中激活语音表征,从而改变听觉感知,这一点可以通过 N1 听觉诱发电位的变化来证明。这一发现支持了视听整合是通过视觉网络影响听觉皮层中的语音表征的理论。我们相信,这将导致关于跨模态映射的新假设的产生,特别是它是否通过直接或间接途径(例如,通过多感觉中介)发生。

相似文献

1
Neural Mechanisms Underlying Cross-Modal Phonetic Encoding.
J Neurosci. 2018 Feb 14;38(7):1835-1849. doi: 10.1523/JNEUROSCI.1566-17.2017. Epub 2017 Dec 20.
2
Cross-modal phonetic encoding facilitates the McGurk illusion and phonemic restoration.
J Neurophysiol. 2018 Dec 1;120(6):2988-3000. doi: 10.1152/jn.00262.2018. Epub 2018 Oct 10.
3
Early and late beta-band power reflect audiovisual perception in the McGurk illusion.
J Neurophysiol. 2015 Apr 1;113(7):2342-50. doi: 10.1152/jn.00783.2014. Epub 2015 Jan 7.
4
Neural evidence accounting for interindividual variability of the McGurk illusion.
Neurosci Lett. 2019 Aug 10;707:134322. doi: 10.1016/j.neulet.2019.134322. Epub 2019 Jun 7.
5
Neurophysiological Indices of Audiovisual Speech Processing Reveal a Hierarchy of Multisensory Integration Effects.
J Neurosci. 2021 Jun 9;41(23):4991-5003. doi: 10.1523/JNEUROSCI.0906-20.2021. Epub 2021 Apr 6.
7
Rethinking the Mechanisms Underlying the McGurk Illusion.
Front Hum Neurosci. 2021 Apr 1;15:616049. doi: 10.3389/fnhum.2021.616049. eCollection 2021.
8
Speech-specific audiovisual integration modulates induced theta-band oscillations.
PLoS One. 2019 Jul 16;14(7):e0219744. doi: 10.1371/journal.pone.0219744. eCollection 2019.
9
Electrophysiological evidence for speech-specific audiovisual integration.
Neuropsychologia. 2014 Jan;53:115-21. doi: 10.1016/j.neuropsychologia.2013.11.011. Epub 2013 Nov 27.
10
Theta oscillations reflect conflict processing in the perception of the McGurk illusion.
Eur J Neurosci. 2018 Oct;48(7):2630-2641. doi: 10.1111/ejn.13804. Epub 2018 Jan 18.

引用本文的文献

1
Neural correlates of phonetic categorization under auditory (phoneme) and visual (grapheme) modalities.
Neuroscience. 2025 Jan 26;565:182-191. doi: 10.1016/j.neuroscience.2024.11.079. Epub 2024 Dec 2.
2
Repeatedly experiencing the McGurk effect induces long-lasting changes in auditory speech perception.
Commun Psychol. 2024 Apr 3;2(1):25. doi: 10.1038/s44271-024-00073-w.
3
Cross-Modal Tinnitus Remediation: A Tentative Theoretical Framework.
Brain Sci. 2024 Jan 19;14(1):95. doi: 10.3390/brainsci14010095.
4
The McGurk Illusion: A Default Mechanism of the Auditory System.
Brain Sci. 2023 Mar 19;13(3):510. doi: 10.3390/brainsci13030510.
5
Electrophysiological Examination of Ambient Speech Processing in Children With Cochlear Implants.
J Speech Lang Hear Res. 2022 Sep 12;65(9):3502-3517. doi: 10.1044/2022_JSLHR-22-00004. Epub 2022 Aug 29.
6
Audition controls the flow of visual time during multisensory perception.
iScience. 2022 Jun 26;25(7):104671. doi: 10.1016/j.isci.2022.104671. eCollection 2022 Jul 15.
7
Functional localization of audiovisual speech using near infrared spectroscopy.
Brain Topogr. 2022 Jul;35(4):416-430. doi: 10.1007/s10548-022-00904-1. Epub 2022 Jul 12.
8
Experimental suppression of transcranial magnetic stimulation-electroencephalography sensory potentials.
Hum Brain Mapp. 2022 Dec 1;43(17):5141-5153. doi: 10.1002/hbm.25990. Epub 2022 Jun 30.
9
A structured ICA-based process for removing auditory evoked potentials.
Sci Rep. 2022 Jan 26;12(1):1391. doi: 10.1038/s41598-022-05397-3.
10

本文引用的文献

1
Auditory, Visual and Audiovisual Speech Processing Streams in Superior Temporal Sulcus.
Front Hum Neurosci. 2017 Apr 7;11:174. doi: 10.3389/fnhum.2017.00174. eCollection 2017.
2
Mouth and Voice: A Relationship between Visual and Auditory Preference in the Human Superior Temporal Sulcus.
J Neurosci. 2017 Mar 8;37(10):2697-2708. doi: 10.1523/JNEUROSCI.2914-16.2017. Epub 2017 Feb 8.
3
Skilled musicians are not subject to the McGurk effect.
Sci Rep. 2016 Jul 25;6:30423. doi: 10.1038/srep30423.
4
A link between individual differences in multisensory speech perception and eye movements.
Atten Percept Psychophys. 2015 May;77(4):1333-41. doi: 10.3758/s13414-014-0821-1.
5
Distributed neural representations of phonological features during speech perception.
J Neurosci. 2015 Jan 14;35(2):634-42. doi: 10.1523/JNEUROSCI.2454-14.2015.
6
Early and late beta-band power reflect audiovisual perception in the McGurk illusion.
J Neurophysiol. 2015 Apr 1;113(7):2342-50. doi: 10.1152/jn.00783.2014. Epub 2015 Jan 7.
7
Putative mechanisms mediating tolerance for audiovisual stimulus onset asynchrony.
J Neurophysiol. 2015 Mar 1;113(5):1437-50. doi: 10.1152/jn.00200.2014. Epub 2014 Dec 10.
8
Eluding the illusion? Schizophrenia, dopamine and the McGurk effect.
Front Hum Neurosci. 2014 Aug 5;8:565. doi: 10.3389/fnhum.2014.00565. eCollection 2014.
9
Effect of attentional load on audiovisual speech perception: evidence from ERPs.
Front Psychol. 2014 Jul 15;5:727. doi: 10.3389/fpsyg.2014.00727. eCollection 2014.
10
Distinct cortical locations for integration of audiovisual speech and the McGurk effect.
Front Psychol. 2014 Jun 2;5:534. doi: 10.3389/fpsyg.2014.00534. eCollection 2014.

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验