• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

无声唇读过程中视觉皮层对运动和分类语音特征的同步化

Visual Cortical Entrainment to Motion and Categorical Speech Features during Silent Lipreading.

作者信息

O'Sullivan Aisling E, Crosse Michael J, Di Liberto Giovanni M, Lalor Edmund C

机构信息

School of Engineering, Trinity College DublinDublin, Ireland; Trinity Centre for Bioengineering, Trinity College DublinDublin, Ireland.

Department of Pediatrics and Department of Neuroscience, Albert Einstein College of Medicine Bronx, NY, USA.

出版信息

Front Hum Neurosci. 2017 Jan 11;10:679. doi: 10.3389/fnhum.2016.00679. eCollection 2016.

DOI:10.3389/fnhum.2016.00679
PMID:28123363
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC5225113/
Abstract

Speech is a multisensory percept, comprising an auditory and visual component. While the content and processing pathways of audio speech have been well characterized, the visual component is less well understood. In this work, we expand current methodologies using system identification to introduce a framework that facilitates the study of visual speech in its natural, continuous form. Specifically, we use models based on the unheard acoustic envelope (E), the motion signal (M) and categorical visual speech features (V) to predict EEG activity during silent lipreading. Our results show that each of these models performs similarly at predicting EEG in visual regions and that respective combinations of the individual models (EV, MV, EM and EMV) provide an improved prediction of the neural activity over their constituent models. In comparing these different combinations, we find that the model incorporating all three types of features (EMV) outperforms the individual models, as well as both the EV and MV models, while it performs similarly to the EM model. Importantly, EM does not outperform EV and MV, which, considering the higher dimensionality of the V model, suggests that more data is needed to clarify this finding. Nevertheless, the performance of EMV, and comparisons of the subject performances for the three individual models, provides further evidence to suggest that visual regions are involved in both low-level processing of stimulus dynamics and categorical speech perception. This framework may prove useful for investigating modality-specific processing of visual speech under naturalistic conditions.

摘要

言语是一种多感官感知,包括听觉和视觉成分。虽然音频言语的内容和处理路径已得到充分表征,但视觉成分的理解却较少。在这项工作中,我们扩展了当前使用系统识别的方法,引入了一个有助于以自然、连续形式研究视觉言语的框架。具体而言,我们使用基于未听到的声学包络(E)、运动信号(M)和分类视觉言语特征(V)的模型来预测默读唇读期间的脑电图活动。我们的结果表明,这些模型中的每一个在预测视觉区域的脑电图方面表现相似,并且各个模型的组合(EV、MV、EM和EMV)在预测神经活动方面比其组成模型有所改进。在比较这些不同组合时,我们发现包含所有三种特征类型的模型(EMV)优于单个模型以及EV和MV模型,同时其表现与EM模型相似。重要的是,EM并不优于EV和MV,考虑到V模型的维度更高,这表明需要更多数据来阐明这一发现。尽管如此,EMV的性能以及三个单个模型的受试者表现比较,提供了进一步的证据表明视觉区域参与了刺激动态的低级处理和分类言语感知。这个框架可能被证明对研究自然条件下视觉言语的模态特异性处理有用。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e438/5225113/18dfeff22c84/fnhum-10-00679-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e438/5225113/a2d612c4d244/fnhum-10-00679-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e438/5225113/822f433883e0/fnhum-10-00679-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e438/5225113/18dfeff22c84/fnhum-10-00679-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e438/5225113/a2d612c4d244/fnhum-10-00679-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e438/5225113/822f433883e0/fnhum-10-00679-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e438/5225113/18dfeff22c84/fnhum-10-00679-g0003.jpg

相似文献

1
Visual Cortical Entrainment to Motion and Categorical Speech Features during Silent Lipreading.无声唇读过程中视觉皮层对运动和分类语音特征的同步化
Front Hum Neurosci. 2017 Jan 11;10:679. doi: 10.3389/fnhum.2016.00679. eCollection 2016.
2
A representation of abstract linguistic categories in the visual system underlies successful lipreading.视觉系统中抽象语言类别的表示是成功唇读的基础。
Neuroimage. 2023 Nov 15;282:120391. doi: 10.1016/j.neuroimage.2023.120391. Epub 2023 Sep 25.
3
MEG Activity in Visual and Auditory Cortices Represents Acoustic Speech-Related Information during Silent Lip Reading.大脑磁图活动在安静默读时代表了视觉和听觉皮层中的与声学语音相关的信息。
eNeuro. 2022 Jun 27;9(3). doi: 10.1523/ENEURO.0209-22.2022. Print 2022 May-Jun.
4
Lip-Reading Enables the Brain to Synthesize Auditory Features of Unknown Silent Speech.唇读使大脑能够合成未知静音语音的听觉特征。
J Neurosci. 2020 Jan 29;40(5):1053-1065. doi: 10.1523/JNEUROSCI.1101-19.2019. Epub 2019 Dec 30.
5
A single-band envelope cue as a supplement to speechreading of segmentals: a comparison of auditory versus tactual presentation.单频段包络线索作为对语音片段唇读的补充:听觉呈现与触觉呈现的比较
Ear Hear. 2001 Jun;22(3):225-35. doi: 10.1097/00003446-200106000-00006.
6
Eye Can Hear Clearly Now: Inverse Effectiveness in Natural Audiovisual Speech Processing Relies on Long-Term Crossmodal Temporal Integration.现在眼睛能“听清”了:自然视听言语处理中的反向有效性依赖于长期跨模态时间整合。
J Neurosci. 2016 Sep 21;36(38):9888-95. doi: 10.1523/JNEUROSCI.1396-16.2016.
7
Congruent audiovisual speech enhances auditory attention decoding with EEG.视听语音一致增强了 EEG 对听觉注意力的解码。
J Neural Eng. 2019 Nov 6;16(6):066033. doi: 10.1088/1741-2552/ab4340.
8
The cortical representation of the speech envelope is earlier for audiovisual speech than audio speech.对于视听语音,其语音包络的皮层表征比听觉语音出现得更早。
J Neurophysiol. 2014 Apr;111(7):1400-8. doi: 10.1152/jn.00690.2013. Epub 2014 Jan 8.
9
Visual speech discrimination and identification of natural and synthetic consonant stimuli.视觉语音辨别以及自然和合成辅音刺激的识别。
Front Psychol. 2015 Jul 13;6:878. doi: 10.3389/fpsyg.2015.00878. eCollection 2015.
10
Audio-visual speech perception in schizophrenia: an fMRI study.精神分裂症中的视听言语感知:一项功能磁共振成像研究。
Psychiatry Res. 2001 Feb 28;106(1):1-14. doi: 10.1016/s0925-4927(00)00081-0.

引用本文的文献

1
Phonological representations of auditory and visual speech in the occipito-temporal cortex and beyond.枕颞叶皮层及其他区域中听觉和视觉言语的语音表征。
J Neurosci. 2025 Apr 30. doi: 10.1523/JNEUROSCI.1415-24.2025.
2
EEG of the Dancing Brain: Decoding Sensory, Motor, and Social Processes during Dyadic Dance.舞动大脑的脑电图:解码双人舞蹈中的感觉、运动和社交过程。
J Neurosci. 2025 May 21;45(21):e2372242025. doi: 10.1523/JNEUROSCI.2372-24.2025.
3
Eye Movements in Silent Visual Speech Track Unheard Acoustic Signals and Relate to Hearing Experience.

本文引用的文献

1
Indexing cortical entrainment to natural speech at the phonemic level: Methodological considerations for applied research.在音素层面将皮层夹带与自然语音进行索引:应用研究的方法学考量
Hear Res. 2017 May;348:70-77. doi: 10.1016/j.heares.2017.02.015. Epub 2017 Feb 27.
2
The Multivariate Temporal Response Function (mTRF) Toolbox: A MATLAB Toolbox for Relating Neural Signals to Continuous Stimuli.多元时间响应函数(mTRF)工具箱:一个用于将神经信号与连续刺激相关联的MATLAB工具箱。
Front Hum Neurosci. 2016 Nov 30;10:604. doi: 10.3389/fnhum.2016.00604. eCollection 2016.
3
Eye Can Hear Clearly Now: Inverse Effectiveness in Natural Audiovisual Speech Processing Relies on Long-Term Crossmodal Temporal Integration.
无声视觉言语中的眼动追踪未听到的声学信号并与听觉体验相关。
eNeuro. 2025 Apr 28;12(4). doi: 10.1523/ENEURO.0055-25.2025. Print 2025 Apr.
4
Auditory cortex encodes lipreading information through spatially distributed activity.听觉皮层通过空间分布的活动来编码唇读信息。
Curr Biol. 2024 Sep 9;34(17):4021-4032.e5. doi: 10.1016/j.cub.2024.07.073. Epub 2024 Aug 16.
5
Effects and prediction of cognitive load on encoding model of brain response to auditory and linguistic stimuli in educational multimedia.认知负荷对教育多媒体中大脑对听觉和语言刺激的反应编码模型的影响及预测
Sci Rep. 2024 Apr 21;14(1):9133. doi: 10.1038/s41598-024-59411-x.
6
Combining Cardiovascular and Pupil Features Using k-Nearest Neighbor Classifiers to Assess Task Demand, Social Context, and Sentence Accuracy During Listening.结合心血管和瞳孔特征,使用 k-最近邻分类器评估在倾听过程中的任务需求、社会环境和句子准确性。
Trends Hear. 2024 Jan-Dec;28:23312165241232551. doi: 10.1177/23312165241232551.
7
Lipreading a naturalistic narrative in a female population: Neural characteristics shared with listening and reading.女性自然叙事唇读:与听读共享的神经特征。
Brain Behav. 2023 Feb;13(2):e2869. doi: 10.1002/brb3.2869. Epub 2022 Dec 29.
8
Differential Auditory and Visual Phase-Locking Are Observed during Audio-Visual Benefit and Silent Lip-Reading for Speech Perception.在听觉-视觉获益和默读唇语感知语音时,观察到听觉和视觉相位锁定的差异。
J Neurosci. 2022 Aug 3;42(31):6108-6120. doi: 10.1523/JNEUROSCI.2476-21.2022. Epub 2022 Jun 27.
9
MEG Activity in Visual and Auditory Cortices Represents Acoustic Speech-Related Information during Silent Lip Reading.大脑磁图活动在安静默读时代表了视觉和听觉皮层中的与声学语音相关的信息。
eNeuro. 2022 Jun 27;9(3). doi: 10.1523/ENEURO.0209-22.2022. Print 2022 May-Jun.
10
Editorial: Neural Tracking: Closing the Gap Between Neurophysiology and Translational Medicine.社论:神经追踪:弥合神经生理学与转化医学之间的差距
Front Neurosci. 2022 Mar 16;16:872600. doi: 10.3389/fnins.2022.872600. eCollection 2022.
现在眼睛能“听清”了:自然视听言语处理中的反向有效性依赖于长期跨模态时间整合。
J Neurosci. 2016 Sep 21;36(38):9888-95. doi: 10.1523/JNEUROSCI.1396-16.2016.
4
Lip movements entrain the observers' low-frequency brain oscillations to facilitate speech intelligibility.唇部动作会带动观察者的低频脑电波振荡,以提高言语清晰度。
Elife. 2016 May 5;5:e14521. doi: 10.7554/eLife.14521.
5
Congruent Visual Speech Enhances Cortical Entrainment to Continuous Auditory Speech in Noise-Free Conditions.在无噪声条件下,匹配的视觉语音增强了皮质对连续听觉语音的同步化。
J Neurosci. 2015 Oct 21;35(42):14195-204. doi: 10.1523/JNEUROSCI.1829-15.2015.
6
Low-Frequency Cortical Entrainment to Speech Reflects Phoneme-Level Processing.对语音的低频皮层夹带反映音素水平的加工。
Curr Biol. 2015 Oct 5;25(19):2457-65. doi: 10.1016/j.cub.2015.08.030. Epub 2015 Sep 24.
7
Prediction and constraint in audiovisual speech perception.视听言语感知中的预测与约束
Cortex. 2015 Jul;68:169-81. doi: 10.1016/j.cortex.2015.03.006. Epub 2015 Mar 20.
8
Neural pathways for visual speech perception.视觉言语感知的神经通路。
Front Neurosci. 2014 Dec 1;8:386. doi: 10.3389/fnins.2014.00386. eCollection 2014.
9
No, there is no 150 ms lead of visual speech on auditory speech, but a range of audiovisual asynchronies varying from small audio lead to large audio lag.不,听觉语音不会领先视觉语音 150 毫秒,而是存在一个从小到大声学提前到大的听觉滞后的视听时滞范围。
PLoS Comput Biol. 2014 Jul 31;10(7):e1003743. doi: 10.1371/journal.pcbi.1003743. eCollection 2014 Jul.
10
Electrocorticography Reveals Enhanced Visual Cortex Responses to Visual Speech.皮层脑电图显示视觉皮层对视觉言语的反应增强。
Cereb Cortex. 2015 Nov;25(11):4103-10. doi: 10.1093/cercor/bhu127. Epub 2014 Jun 5.