• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

在真实声学场景中对注意到的语音进行抗噪皮层追踪。

Noise-robust cortical tracking of attended speech in real-world acoustic scenes.

作者信息

Fuglsang Søren Asp, Dau Torsten, Hjortkjær Jens

机构信息

Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, Ørsteds Plads, Building 352, 2800 Kgs. Lyngby, Denmark.

Hearing Systems Group, Department of Electrical Engineering, Technical University of Denmark, Ørsteds Plads, Building 352, 2800 Kgs. Lyngby, Denmark.

出版信息

Neuroimage. 2017 Aug 1;156:435-444. doi: 10.1016/j.neuroimage.2017.04.026. Epub 2017 Apr 13.

DOI:10.1016/j.neuroimage.2017.04.026
PMID:28412441
Abstract

Selectively attending to one speaker in a multi-speaker scenario is thought to synchronize low-frequency cortical activity to the attended speech signal. In recent studies, reconstruction of speech from single-trial electroencephalogram (EEG) data has been used to decode which talker a listener is attending to in a two-talker situation. It is currently unclear how this generalizes to more complex sound environments. Behaviorally, speech perception is robust to the acoustic distortions that listeners typically encounter in everyday life, but it is unknown whether this is mirrored by a noise-robust neural tracking of attended speech. Here we used advanced acoustic simulations to recreate real-world acoustic scenes in the laboratory. In virtual acoustic realities with varying amounts of reverberation and number of interfering talkers, listeners selectively attended to the speech stream of a particular talker. Across the different listening environments, we found that the attended talker could be accurately decoded from single-trial EEG data irrespective of the different distortions in the acoustic input. For highly reverberant environments, speech envelopes reconstructed from neural responses to the distorted stimuli resembled the original clean signal more than the distorted input. With reverberant speech, we observed a late cortical response to the attended speech stream that encoded temporal modulations in the speech signal without its reverberant distortion. Single-trial attention decoding accuracies based on 40-50s long blocks of data from 64 scalp electrodes were equally high (80-90% correct) in all considered listening environments and remained statistically significant using down to 10 scalp electrodes and short (<30-s) unaveraged EEG segments. In contrast to the robust decoding of the attended talker we found that decoding of the unattended talker deteriorated with the acoustic distortions. These results suggest that cortical activity tracks an attended speech signal in a way that is invariant to acoustic distortions encountered in real-life sound environments. Noise-robust attention decoding additionally suggests a potential utility of stimulus reconstruction techniques in attention-controlled brain-computer interfaces.

摘要

在多说话者场景中选择性地关注一个说话者被认为会使低频皮层活动与被关注的语音信号同步。在最近的研究中,利用单次试验脑电图(EEG)数据进行语音重构已被用于解码在双说话者情境中听众正在关注哪个说话者。目前尚不清楚这如何推广到更复杂的声音环境。从行为上来说,语音感知对听众在日常生活中通常遇到的声学失真具有鲁棒性,但尚不清楚这是否通过对被关注语音的噪声鲁棒神经跟踪得到反映。在这里,我们使用先进的声学模拟在实验室中重现真实世界的声学场景。在具有不同混响量和干扰说话者数量的虚拟声学环境中,听众选择性地关注特定说话者的语音流。在不同的聆听环境中,我们发现无论声学输入中的不同失真情况如何,都可以从单次试验EEG数据中准确解码出被关注的说话者。对于高混响环境,从对失真刺激的神经反应重构的语音包络比失真输入更类似于原始纯净信号。对于混响语音,我们观察到对被关注语音流的晚期皮层反应,该反应对语音信号中的时间调制进行编码而不包含其混响失真。基于来自64个头皮电极的40 - 50秒长数据块的单次试验注意力解码准确率在所有考虑的聆听环境中同样很高(正确率为80 - 90%),并且使用低至10个头皮电极和短(<30秒)未平均的EEG片段时仍具有统计学意义。与对被关注说话者的鲁棒解码形成对比的是,我们发现对未被关注说话者的解码随着声学失真而恶化。这些结果表明,皮层活动以一种对现实生活声音环境中遇到的声学失真不变的方式跟踪被关注的语音信号。噪声鲁棒注意力解码还表明刺激重构技术在注意力控制脑机接口中的潜在效用。

相似文献

1
Noise-robust cortical tracking of attended speech in real-world acoustic scenes.在真实声学场景中对注意到的语音进行抗噪皮层追踪。
Neuroimage. 2017 Aug 1;156:435-444. doi: 10.1016/j.neuroimage.2017.04.026. Epub 2017 Apr 13.
2
Effects of Sensorineural Hearing Loss on Cortical Synchronization to Competing Speech during Selective Attention.感音神经性听力损失对选择性注意期间竞争语音的皮层同步的影响。
J Neurosci. 2020 Mar 18;40(12):2562-2572. doi: 10.1523/JNEUROSCI.1936-19.2020. Epub 2020 Feb 24.
3
EEG decoding of the target speaker in a cocktail party scenario: considerations regarding dynamic switching of talker location.鸡尾酒会场景中目标说话人的 EEG 解码:关于说话人位置动态切换的考虑因素。
J Neural Eng. 2019 Jun;16(3):036017. doi: 10.1088/1741-2552/ab0cf1. Epub 2019 Mar 5.
4
Where is the cocktail party? Decoding locations of attended and unattended moving sound sources using EEG.鸡尾酒会在哪里?使用 EEG 解码有注意和无注意移动声源的位置。
Neuroimage. 2020 Jan 15;205:116283. doi: 10.1016/j.neuroimage.2019.116283. Epub 2019 Oct 17.
5
Impact of Different Acoustic Components on EEG-Based Auditory Attention Decoding in Noisy and Reverberant Conditions.不同声成分对噪声和混响环境下基于 EEG 的听觉注意解码的影响。
IEEE Trans Neural Syst Rehabil Eng. 2019 Apr;27(4):652-663. doi: 10.1109/TNSRE.2019.2903404. Epub 2019 Mar 7.
6
Congruent audiovisual speech enhances auditory attention decoding with EEG.视听语音一致增强了 EEG 对听觉注意力的解码。
J Neural Eng. 2019 Nov 6;16(6):066033. doi: 10.1088/1741-2552/ab4340.
7
Neural Speech Tracking in the Theta and in the Delta Frequency Band Differentially Encode Clarity and Comprehension of Speech in Noise.在θ和δ频段的神经语音跟踪对语音清晰度和噪声中理解的编码有差异。
J Neurosci. 2019 Jul 17;39(29):5750-5759. doi: 10.1523/JNEUROSCI.1828-18.2019. Epub 2019 May 20.
8
Selective cortical representation of attended speaker in multi-talker speech perception.选择性皮层对多说话人语音感知中被注意说话人的代表。
Nature. 2012 May 10;485(7397):233-6. doi: 10.1038/nature11020.
9
Left Superior Temporal Gyrus Is Coupled to Attended Speech in a Cocktail-Party Auditory Scene.左颞上回在鸡尾酒会听觉场景中与被关注的语音相关联。
J Neurosci. 2016 Feb 3;36(5):1596-606. doi: 10.1523/JNEUROSCI.1730-15.2016.
10
Look at me when I'm talking to you: Selective attention at a multisensory cocktail party can be decoded using stimulus reconstruction and alpha power modulations.当我和你说话时看着我:使用刺激重建和阿尔法功率调制,可以对多感官鸡尾酒会上的选择性注意进行解码。
Eur J Neurosci. 2019 Oct;50(8):3282-3295. doi: 10.1111/ejn.14425. Epub 2019 May 17.

引用本文的文献

1
Age-Related Differences in EEG-Based Speech Reception Threshold Estimation Using Scalp and Ear-EEG.使用头皮脑电图和耳部脑电图基于脑电图的言语接受阈值估计中的年龄相关差异。
Trends Hear. 2025 Jan-Dec;29:23312165251372462. doi: 10.1177/23312165251372462. Epub 2025 Aug 25.
2
Linear modeling of brain activity during selective attention to continuous speech: the critical role of the N1 effect in event-related potentials to acoustic edges.对连续语音进行选择性注意时大脑活动的线性建模:N1效应在与听觉边缘相关的事件相关电位中的关键作用。
Cogn Neurodyn. 2025 Dec;19(1):110. doi: 10.1007/s11571-025-10289-z. Epub 2025 Jul 2.
3
Neural Speech Tracking during Selective Attention: A Spatially Realistic Audiovisual Study.
选择性注意期间的神经语音追踪:一项空间逼真的视听研究。
eNeuro. 2025 Jun 24;12(6). doi: 10.1523/ENEURO.0132-24.2025. Print 2025 Jun.
4
From beeps to streets: unveiling sensory input and relevance across auditory contexts.从蜂鸣声到街道:揭示跨听觉情境的感官输入及相关性。
Front Neuroergon. 2025 Apr 28;6:1571356. doi: 10.3389/fnrgo.2025.1571356. eCollection 2025.
5
Speech Reception Threshold Estimation via EEG-Based Continuous Speech Envelope Reconstruction.基于脑电图的连续语音包络重构的言语接受阈值估计
Eur J Neurosci. 2025 Mar;61(6):e70083. doi: 10.1111/ejn.70083.
6
Improving Tracking of Selective Attention in Hearing Aid Users: The Role of Noise Reduction and Nonlinearity Compensation.改善助听器使用者选择性注意力的追踪:降噪与非线性补偿的作用。
eNeuro. 2025 Feb 19;12(2). doi: 10.1523/ENEURO.0275-24.2025. Print 2025 Feb.
7
Adaptation to Reverberation for Speech Perception: A Systematic Review.适应混响的语音感知:系统评价。
Trends Hear. 2024 Jan-Dec;28:23312165241273399. doi: 10.1177/23312165241273399.
8
Neural Decoding of the Speech Envelope: Effects of Intelligibility and Spectral Degradation.语音包络的神经解码:可懂度和频谱劣化的影响。
Trends Hear. 2024 Jan-Dec;28:23312165241266316. doi: 10.1177/23312165241266316.
9
Sensory and Perceptual Decisional Processes Underlying the Perception of Reverberant Auditory Environments.基于混响听觉环境感知的感觉和知觉决策过程。
eNeuro. 2024 Aug 20;11(8). doi: 10.1523/ENEURO.0122-24.2024. Print 2024 Aug.
10
Convolutional neural networks can identify brain interactions involved in decoding spatial auditory attention.卷积神经网络可以识别解码空间听觉注意力涉及的大脑交互。
PLoS Comput Biol. 2024 Aug 8;20(8):e1012376. doi: 10.1371/journal.pcbi.1012376. eCollection 2024 Aug.