• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于稳健脑电图的嘈杂环境下高均方根电平语音片段听觉注意力解码

Robust EEG-Based Decoding of Auditory Attention With High-RMS-Level Speech Segments in Noisy Conditions.

作者信息

Wang Lei, Wu Ed X, Chen Fei

机构信息

Department of Electrical and Electronic Engineering, Southern University of Science and Technology, Shenzhen, China.

Department of Electrical and Electronic Engineering, The University of Hong Kong, Hong Kong, Hong Kong.

出版信息

Front Hum Neurosci. 2020 Oct 7;14:557534. doi: 10.3389/fnhum.2020.557534. eCollection 2020.

DOI:10.3389/fnhum.2020.557534
PMID:33132874
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7576187/
Abstract

The attended speech stream can be detected robustly, even in adverse auditory scenarios with auditory attentional modulation, and can be decoded using electroencephalographic (EEG) data. Speech segmentation based on the relative root-mean-square (RMS) intensity can be used to estimate segmental contributions to perception in noisy conditions. High-RMS-level segments contain crucial information for speech perception. Hence, this study aimed to investigate the effect of high-RMS-level speech segments on auditory attention decoding performance under various signal-to-noise ratio (SNR) conditions. Scalp EEG signals were recorded when subjects listened to the attended speech stream in the mixed speech narrated concurrently by two Mandarin speakers. The temporal response function was used to identify the attended speech from EEG responses of tracking to the temporal envelopes of intact speech and high-RMS-level speech segments alone, respectively. Auditory decoding performance was then analyzed under various SNR conditions by comparing EEG correlations to the attended and ignored speech streams. The accuracy of auditory attention decoding based on the temporal envelope with high-RMS-level speech segments was not inferior to that based on the temporal envelope of intact speech. Cortical activity correlated more strongly with attended than with ignored speech under different SNR conditions. These results suggest that EEG recordings corresponding to high-RMS-level speech segments carry crucial information for the identification and tracking of attended speech in the presence of background noise. This study also showed that with the modulation of auditory attention, attended speech can be decoded more robustly from neural activity than from behavioral measures under a wide range of SNR.

摘要

即使在存在听觉注意调制的不利听觉场景中,被关注的语音流也能被可靠地检测到,并且可以使用脑电图(EEG)数据进行解码。基于相对均方根(RMS)强度的语音分割可用于估计在噪声环境中各语音段对语音感知的贡献。高RMS水平的语音段包含语音感知的关键信息。因此,本研究旨在探讨在各种信噪比(SNR)条件下,高RMS水平语音段对听觉注意解码性能的影响。当受试者在两名说普通话的人同时叙述的混合语音中收听被关注的语音流时,记录头皮EEG信号。时间响应函数分别用于从仅跟踪完整语音和高RMS水平语音段的时间包络的EEG响应中识别被关注的语音。然后,通过比较EEG与被关注和被忽略语音流的相关性,分析在各种SNR条件下的听觉解码性能。基于高RMS水平语音段时间包络的听觉注意解码准确率不低于基于完整语音时间包络的解码准确率。在不同SNR条件下,皮层活动与被关注语音的相关性比与被忽略语音的相关性更强。这些结果表明,在存在背景噪声的情况下,与高RMS水平语音段对应的EEG记录携带了用于识别和跟踪被关注语音的关键信息。本研究还表明,在听觉注意的调制下,在广泛的SNR范围内,从神经活动中解码被关注语音比从行为测量中更稳健。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/71a1/7576187/466d9102dd02/fnhum-14-557534-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/71a1/7576187/22ab39370158/fnhum-14-557534-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/71a1/7576187/dbdc6716a196/fnhum-14-557534-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/71a1/7576187/0b8aee3cde43/fnhum-14-557534-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/71a1/7576187/fcbe7c400f54/fnhum-14-557534-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/71a1/7576187/7b36832c0bfb/fnhum-14-557534-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/71a1/7576187/466d9102dd02/fnhum-14-557534-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/71a1/7576187/22ab39370158/fnhum-14-557534-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/71a1/7576187/dbdc6716a196/fnhum-14-557534-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/71a1/7576187/0b8aee3cde43/fnhum-14-557534-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/71a1/7576187/fcbe7c400f54/fnhum-14-557534-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/71a1/7576187/7b36832c0bfb/fnhum-14-557534-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/71a1/7576187/466d9102dd02/fnhum-14-557534-g006.jpg

相似文献

1
Robust EEG-Based Decoding of Auditory Attention With High-RMS-Level Speech Segments in Noisy Conditions.基于稳健脑电图的嘈杂环境下高均方根电平语音片段听觉注意力解码
Front Hum Neurosci. 2020 Oct 7;14:557534. doi: 10.3389/fnhum.2020.557534. eCollection 2020.
2
EEG-based auditory attention decoding using speech-level-based segmented computational models.基于脑电的听觉注意解码,使用基于语音分段的计算模型。
J Neural Eng. 2021 May 25;18(4). doi: 10.1088/1741-2552/abfeba.
3
A Speech-Level-Based Segmented Model to Decode the Dynamic Auditory Attention States in the Competing Speaker Scenes.一种基于语音水平的分段模型,用于解码竞争说话者场景中的动态听觉注意状态。
Front Neurosci. 2022 Feb 10;15:760611. doi: 10.3389/fnins.2021.760611. eCollection 2021.
4
Cortical auditory responses index the contributions of different RMS-level-dependent segments to speech intelligibility.皮层听觉反应指数不同 RMS 水平相关段对口音可懂度的贡献。
Hear Res. 2019 Nov;383:107808. doi: 10.1016/j.heares.2019.107808. Epub 2019 Oct 4.
5
Noise-robust cortical tracking of attended speech in real-world acoustic scenes.在真实声学场景中对注意到的语音进行抗噪皮层追踪。
Neuroimage. 2017 Aug 1;156:435-444. doi: 10.1016/j.neuroimage.2017.04.026. Epub 2017 Apr 13.
6
Robust decoding of the speech envelope from EEG recordings through deep neural networks.通过深度神经网络从 EEG 记录中稳健地解码语音包络。
J Neural Eng. 2022 Jul 6;19(4). doi: 10.1088/1741-2552/ac7976.
7
Impact of Different Acoustic Components on EEG-Based Auditory Attention Decoding in Noisy and Reverberant Conditions.不同声成分对噪声和混响环境下基于 EEG 的听觉注意解码的影响。
IEEE Trans Neural Syst Rehabil Eng. 2019 Apr;27(4):652-663. doi: 10.1109/TNSRE.2019.2903404. Epub 2019 Mar 7.
8
Effects of Sensorineural Hearing Loss on Cortical Synchronization to Competing Speech during Selective Attention.感音神经性听力损失对选择性注意期间竞争语音的皮层同步的影响。
J Neurosci. 2020 Mar 18;40(12):2562-2572. doi: 10.1523/JNEUROSCI.1936-19.2020. Epub 2020 Feb 24.
9
EEG-based auditory attention detection: boundary conditions for background noise and speaker positions.基于脑电图的听觉注意力检测:背景噪声和说话人位置的边界条件。
J Neural Eng. 2018 Dec;15(6):066017. doi: 10.1088/1741-2552/aae0a6. Epub 2018 Sep 12.
10
The Right Temporoparietal Junction Supports Speech Tracking During Selective Listening: Evidence from Concurrent EEG-fMRI.右侧颞顶联合区在选择性倾听过程中支持言语追踪:来自同步脑电图-功能磁共振成像的证据。
J Neurosci. 2017 Nov 22;37(47):11505-11516. doi: 10.1523/JNEUROSCI.1007-17.2017. Epub 2017 Oct 23.

引用本文的文献

1
A Brain-Computer Interface for Improving Auditory Attention in Multi-Talker Environments.一种用于在多说话者环境中改善听觉注意力的脑机接口。
bioRxiv. 2025 Mar 13:2025.03.13.641661. doi: 10.1101/2025.03.13.641661.
2
A Speech-Level-Based Segmented Model to Decode the Dynamic Auditory Attention States in the Competing Speaker Scenes.一种基于语音水平的分段模型,用于解码竞争说话者场景中的动态听觉注意状态。
Front Neurosci. 2022 Feb 10;15:760611. doi: 10.3389/fnins.2021.760611. eCollection 2021.
3
Auditory Attention Detection via Cross-Modal Attention.

本文引用的文献

1
Effect of Task and Attention on Neural Tracking of Speech.任务和注意力对语音神经追踪的影响。
Front Neurosci. 2019 Sep 16;13:977. doi: 10.3389/fnins.2019.00977. eCollection 2019.
2
Cortical auditory responses index the contributions of different RMS-level-dependent segments to speech intelligibility.皮层听觉反应指数不同 RMS 水平相关段对口音可懂度的贡献。
Hear Res. 2019 Nov;383:107808. doi: 10.1016/j.heares.2019.107808. Epub 2019 Oct 4.
3
Neural Entrainment and Attentional Selection in the Listening Brain.听脑中的神经同步与注意选择。
通过跨模态注意力进行听觉注意力检测
Front Neurosci. 2021 Jul 21;15:652058. doi: 10.3389/fnins.2021.652058. eCollection 2021.
Trends Cogn Sci. 2019 Nov;23(11):913-926. doi: 10.1016/j.tics.2019.08.004. Epub 2019 Oct 9.
4
Speech fine structure contains critical temporal cues to support speech segmentation.言语的细微结构包含支持言语分段的关键时间线索。
Neuroimage. 2019 Nov 15;202:116152. doi: 10.1016/j.neuroimage.2019.116152. Epub 2019 Sep 1.
5
Factors affecting the intelligibility of high-intensity-level-based speech.影响高强度级语音可懂度的因素。
J Acoust Soc Am. 2019 Aug;146(2):EL151. doi: 10.1121/1.5122190.
6
Semantic Context Enhances the Early Auditory Encoding of Natural Speech.语义语境增强了对自然语音的早期听觉编码。
J Neurosci. 2019 Sep 18;39(38):7564-7575. doi: 10.1523/JNEUROSCI.0584-19.2019. Epub 2019 Aug 1.
7
Cortical Tracking of Complex Sound Envelopes: Modeling the Changes in Response with Intensity.皮层对复杂声音包络的追踪:随强度变化的响应建模。
eNeuro. 2019 Jun 27;6(3). doi: 10.1523/ENEURO.0082-19.2019. Print 2019 May/Jun.
8
A Tutorial on Auditory Attention Identification Methods.听觉注意力识别方法教程。
Front Neurosci. 2019 Mar 19;13:153. doi: 10.3389/fnins.2019.00153. eCollection 2019.
9
Impact of Different Acoustic Components on EEG-Based Auditory Attention Decoding in Noisy and Reverberant Conditions.不同声成分对噪声和混响环境下基于 EEG 的听觉注意解码的影响。
IEEE Trans Neural Syst Rehabil Eng. 2019 Apr;27(4):652-663. doi: 10.1109/TNSRE.2019.2903404. Epub 2019 Mar 7.
10
EEG decoding of the target speaker in a cocktail party scenario: considerations regarding dynamic switching of talker location.鸡尾酒会场景中目标说话人的 EEG 解码:关于说话人位置动态切换的考虑因素。
J Neural Eng. 2019 Jun;16(3):036017. doi: 10.1088/1741-2552/ab0cf1. Epub 2019 Mar 5.