• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

相似文献

1
Vision perceptually restores auditory spectral dynamics in speech.视觉在感知上恢复了言语中的听觉频谱动态。
Proc Natl Acad Sci U S A. 2020 Jul 21;117(29):16920-16927. doi: 10.1073/pnas.2002887117. Epub 2020 Jul 6.
2
Neurophysiological Indices of Audiovisual Speech Processing Reveal a Hierarchy of Multisensory Integration Effects.神经生理指标揭示视听言语加工中的多感觉整合效应层次结构。
J Neurosci. 2021 Jun 9;41(23):4991-5003. doi: 10.1523/JNEUROSCI.0906-20.2021. Epub 2021 Apr 6.
3
The use of visible speech cues for improving auditory detection of spoken sentences.使用可见语音线索来提高对口语句子的听觉检测。
J Acoust Soc Am. 2000 Sep;108(3 Pt 1):1197-208. doi: 10.1121/1.1288668.
4
Speech identification in noise: Contribution of temporal, spectral, and visual speech cues.噪声中的语音识别:时域、频域和可见语音线索的贡献。
J Acoust Soc Am. 2009 Dec;126(6):3246-57. doi: 10.1121/1.3250425.
5
Neural Mechanisms Underlying Cross-Modal Phonetic Encoding.跨模态语音编码的神经机制。
J Neurosci. 2018 Feb 14;38(7):1835-1849. doi: 10.1523/JNEUROSCI.1566-17.2017. Epub 2017 Dec 20.
6
How visual cues to speech rate influence speech perception.言语速度的视觉提示如何影响言语感知。
Q J Exp Psychol (Hove). 2020 Oct;73(10):1523-1536. doi: 10.1177/1747021820914564. Epub 2020 Apr 20.
7
Timing in audiovisual speech perception: A mini review and new psychophysical data.视听言语感知中的时间因素:一篇小型综述及新的心理物理学数据
Atten Percept Psychophys. 2016 Feb;78(2):583-601. doi: 10.3758/s13414-015-1026-y.
8
Visual speech information: a help or hindrance in perceptual processing of dysarthric speech.视觉言语信息:在构音障碍语音的感知处理中是助力还是阻碍?
J Acoust Soc Am. 2015 Mar;137(3):1473-80. doi: 10.1121/1.4913770.
9
Assessment of Spectral and Temporal Resolution in Cochlear Implant Users Using Psychoacoustic Discrimination and Speech Cue Categorization.使用心理声学辨别和语音线索分类评估人工耳蜗使用者的频谱和时间分辨率
Ear Hear. 2016 Nov/Dec;37(6):e377-e390. doi: 10.1097/AUD.0000000000000328.
10
Low- and high-frequency cortical brain oscillations reflect dissociable mechanisms of concurrent speech segregation in noise.低频和高频皮层脑振荡反映了噪声中并行语音分离的不同机制。
Hear Res. 2018 Apr;361:92-102. doi: 10.1016/j.heares.2018.01.006. Epub 2018 Feb 2.

引用本文的文献

1
Minimal Differences in Auditory and Visual Oddball Tasks in Autism: A Systematic Review and Meta-Analysis.自闭症患者在听觉和视觉Oddball任务中的微小差异:系统评价与荟萃分析
J Autism Dev Disord. 2025 Mar 8. doi: 10.1007/s10803-025-06772-5.
2
Brain Encoding of Naturalistic, Continuous, and Unpredictable Tactile Events.自然连续不可预测触觉事件的大脑编码。
eNeuro. 2024 Sep 24;11(9). doi: 10.1523/ENEURO.0238-24.2024. Print 2024 Sep.
3
Evidence for a Causal Dissociation of the McGurk Effect and Congruent Audiovisual Speech Perception via TMS to the Left pSTS.经左颞上沟重复经颅磁刺激对 McGurk 效应和一致视听言语感知的因果分离的证据。
Multisens Res. 2024 Aug 16;37(4-5):341-363. doi: 10.1163/22134808-bja10129.
4
Auditory cortex encodes lipreading information through spatially distributed activity.听觉皮层通过空间分布的活动来编码唇读信息。
Curr Biol. 2024 Sep 9;34(17):4021-4032.e5. doi: 10.1016/j.cub.2024.07.073. Epub 2024 Aug 16.
5
Multisensory and lexical information in speech perception.言语感知中的多感官和词汇信息。
Front Hum Neurosci. 2024 Jan 8;17:1331129. doi: 10.3389/fnhum.2023.1331129. eCollection 2023.
6
Evidence for a Causal Dissociation of the McGurk Effect and Congruent Audiovisual Speech Perception via TMS.通过经颅磁刺激(TMS)证明麦格克效应与一致的视听语音感知之间存在因果分离。
bioRxiv. 2023 Nov 27:2023.11.27.568892. doi: 10.1101/2023.11.27.568892.
7
The Development of Speaking and Singing in Infants May Play a Role in Genomics and Dementia in Humans.婴儿语言和歌唱能力的发展可能在人类基因组学和痴呆症中发挥作用。
Brain Sci. 2023 Aug 11;13(8):1190. doi: 10.3390/brainsci13081190.
8
Neurosensory development of the four brainstem-projecting sensory systems and their integration in the telencephalon.脑干投射感觉系统的四个神经系统感觉发育及其在端脑的整合。
Front Neural Circuits. 2022 Sep 23;16:913480. doi: 10.3389/fncir.2022.913480. eCollection 2022.
9
MEG Activity in Visual and Auditory Cortices Represents Acoustic Speech-Related Information during Silent Lip Reading.大脑磁图活动在安静默读时代表了视觉和听觉皮层中的与声学语音相关的信息。
eNeuro. 2022 Jun 27;9(3). doi: 10.1523/ENEURO.0209-22.2022. Print 2022 May-Jun.
10
Cortical tracking of formant modulations derived from silently presented lip movements and its decline with age.基于无声呈现的嘴唇运动的共振峰调制的皮层追踪及其随年龄的衰退。
Cereb Cortex. 2022 Oct 20;32(21):4818-4833. doi: 10.1093/cercor/bhab518.

本文引用的文献

1
[i] is Lighter and More Greenish Than [o]: Intrinsic Association Between Vowel Sounds and Colors.[i] 比 [o] 更浅且更偏绿色:元音与颜色之间的内在关联。
Multisens Res. 2018 Jan 1;31(5):419-437. doi: 10.1163/22134808-00002581.
2
Hearing-impaired listeners show increased audiovisual benefit when listening to speech in noise.听障人士在噪声环境下聆听言语时,会获得更大的视听增益。
Neuroimage. 2019 Aug 1;196:261-268. doi: 10.1016/j.neuroimage.2019.04.017. Epub 2019 Apr 9.
3
Audiovisual speech perception: A new approach and implications for clinical populations.视听言语感知:一种新方法及其对临床人群的启示。
Lang Linguist Compass. 2017 Mar;11(3):77-91. doi: 10.1111/lnc3.12237. Epub 2017 Mar 26.
4
The coupling between auditory and motor cortices is rate-restricted: Evidence for an intrinsic speech-motor rhythm.听觉皮层和运动皮层之间的耦合受到频率限制:固有言语-运动节律的证据。
Sci Adv. 2018 Feb 7;4(2):eaao3842. doi: 10.1126/sciadv.aao3842. eCollection 2018 Feb.
5
A Causal Inference Model Explains Perception of the McGurk Effect and Other Incongruent Audiovisual Speech.一种因果推理模型解释了麦格克效应及其他不一致视听言语的感知。
PLoS Comput Biol. 2017 Feb 16;13(2):e1005229. doi: 10.1371/journal.pcbi.1005229. eCollection 2017 Feb.
6
Mouth and Voice: A Relationship between Visual and Auditory Preference in the Human Superior Temporal Sulcus.嘴巴与声音:人类颞上沟中视觉与听觉偏好之间的关系。
J Neurosci. 2017 Mar 8;37(10):2697-2708. doi: 10.1523/JNEUROSCI.2914-16.2017. Epub 2017 Feb 8.
7
Rapid tuning shifts in human auditory cortex enhance speech intelligibility.快速调谐人听觉皮层的转变增强了言语可懂度。
Nat Commun. 2016 Dec 20;7:13654. doi: 10.1038/ncomms13654.
8
The auditory representation of speech sounds in human motor cortex.人类运动皮层中语音的听觉表征。
Elife. 2016 Mar 4;5:e12577. doi: 10.7554/eLife.12577.
9
Human Superior Temporal Gyrus Organization of Spectrotemporal Modulation Tuning Derived from Speech Stimuli.源自语音刺激的人类颞上回频谱时间调制调谐组织
J Neurosci. 2016 Feb 10;36(6):2014-26. doi: 10.1523/JNEUROSCI.1779-15.2016.
10
No evidence of somatotopic place of articulation feature mapping in motor cortex during passive speech perception.在被动言语感知过程中,运动皮层未发现发音部位特征映射的躯体定位证据。
Psychon Bull Rev. 2016 Aug;23(4):1231-40. doi: 10.3758/s13423-015-0988-z.

视觉在感知上恢复了言语中的听觉频谱动态。

Vision perceptually restores auditory spectral dynamics in speech.

机构信息

Department of Psychology, University of Michigan, Ann Arbor, MI 48109;

Department of Psychology, Northwestern University, Evanston, IL 60208.

出版信息

Proc Natl Acad Sci U S A. 2020 Jul 21;117(29):16920-16927. doi: 10.1073/pnas.2002887117. Epub 2020 Jul 6.

DOI:10.1073/pnas.2002887117
PMID:32632010
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7382243/
Abstract

Visual speech facilitates auditory speech perception, but the visual cues responsible for these benefits and the information they provide remain unclear. Low-level models emphasize basic temporal cues provided by mouth movements, but these impoverished signals may not fully account for the richness of auditory information provided by visual speech. High-level models posit interactions among abstract categorical (i.e., phonemes/visemes) or amodal (e.g., articulatory) speech representations, but require lossy remapping of speech signals onto abstracted representations. Because visible articulators shape the spectral content of speech, we hypothesized that the perceptual system might exploit natural correlations between midlevel visual (oral deformations) and auditory speech features (frequency modulations) to extract detailed spectrotemporal information from visual speech without employing high-level abstractions. Consistent with this hypothesis, we found that the time-frequency dynamics of oral resonances (formants) could be predicted with unexpectedly high precision from the changing shape of the mouth during speech. When isolated from other speech cues, speech-based shape deformations improved perceptual sensitivity for corresponding frequency modulations, suggesting that listeners could exploit this cross-modal correspondence to facilitate perception. To test whether this type of correspondence could improve speech comprehension, we selectively degraded the spectral or temporal dimensions of auditory sentence spectrograms to assess how well visual speech facilitated comprehension under each degradation condition. Visual speech produced drastically larger enhancements during spectral degradation, suggesting a condition-specific facilitation effect driven by cross-modal recovery of auditory speech spectra. The perceptual system may therefore use audiovisual correlations rooted in oral acoustics to extract detailed spectrotemporal information from visual speech.

摘要

视觉语音促进听觉语音感知,但负责这些益处的视觉线索以及它们提供的信息仍不清楚。低水平模型强调由口部运动提供的基本时间线索,但这些简化的信号可能无法充分说明视觉语音提供的听觉信息的丰富性。高水平模型假设抽象类别(即音素/视位)或无模态(例如,发音)语音表示之间的相互作用,但需要对语音信号进行有损的重新映射到抽象表示上。由于可见的发音器官塑造了语音的频谱内容,我们假设感知系统可能利用中观视觉(口腔变形)和听觉语音特征(频率调制)之间的自然相关性,无需采用高级抽象,从视觉语音中提取详细的时频信息。与该假设一致,我们发现口腔共振(共振峰)的时频动态可以从语音过程中口形的变化以出人意料的高精度来预测。当与其他语音线索隔离开来时,基于语音的形状变形可以提高对相应频率调制的感知灵敏度,这表明听众可以利用这种跨模态对应关系来促进感知。为了测试这种对应关系是否可以改善语音理解,我们选择性地降低了听觉句子声谱图的光谱或时间维度,以评估在每种降解条件下视觉语音对理解的促进程度。在光谱降解期间,视觉语音产生了极大的增强,这表明这是一种由听觉语音频谱的跨模态恢复驱动的特定条件下的促进效应。因此,感知系统可能会使用根植于口腔声学的视听相关性,从视觉语音中提取详细的时频信息。