• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

一种基于语音水平的分段模型,用于解码竞争说话者场景中的动态听觉注意状态。

A Speech-Level-Based Segmented Model to Decode the Dynamic Auditory Attention States in the Competing Speaker Scenes.

作者信息

Wang Lei, Wang Yihan, Liu Zhixing, Wu Ed X, Chen Fei

机构信息

Department of Electrical and Electronic Engineering, Southern University of Science and Technology, Shenzhen, China.

Department of Electrical and Electronic Engineering, The University of Hong Kong, Pokfulam, Hong Kong SAR, China.

出版信息

Front Neurosci. 2022 Feb 10;15:760611. doi: 10.3389/fnins.2021.760611. eCollection 2021.

DOI:10.3389/fnins.2021.760611
PMID:35221885
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8866945/
Abstract

In the competing speaker environments, human listeners need to focus or switch their auditory attention according to dynamic intentions. The reliable cortical tracking ability to the speech envelope is an effective feature for decoding the target speech from the neural signals. Moreover, previous studies revealed that the root mean square (RMS)-level-based speech segmentation made a great contribution to the target speech perception with the modulation of sustained auditory attention. This study further investigated the effect of the RMS-level-based speech segmentation on the auditory attention decoding (AAD) performance with both sustained and switched attention in the competing speaker auditory scenes. Objective biomarkers derived from the cortical activities were also developed to index the dynamic auditory attention states. In the current study, subjects were asked to concentrate or switch their attention between two competing speaker streams. The neural responses to the higher- and lower-RMS-level speech segments were analyzed the linear temporal response function (TRF) before and after the attention switching from one to the other speaker stream. Furthermore, the AAD performance decoded by the unified TRF decoding model was compared to that by the speech-RMS-level-based segmented decoding model with the dynamic change of the auditory attention states. The results showed that the weight of the typical TRF component approximately 100-ms time lag was sensitive to the switching of the auditory attention. Compared to the unified AAD model, the segmented AAD model improved attention decoding performance under both the sustained and switched auditory attention modulations in a wide range of signal-to-masker ratios (SMRs). In the competing speaker scenes, the TRF weight and AAD accuracy could be used as effective indicators to detect the changes of the auditory attention. In addition, with a wide range of SMRs (i.e., from 6 to -6 dB in this study), the segmented AAD model showed the robust decoding performance even with short decision window length, suggesting that this speech-RMS-level-based model has the potential to decode dynamic attention states in the realistic auditory scenarios.

摘要

在竞争性说话者环境中,人类听众需要根据动态意图集中或切换他们的听觉注意力。对语音包络的可靠皮层跟踪能力是从神经信号中解码目标语音的有效特征。此外,先前的研究表明,基于均方根(RMS)电平的语音分割在持续听觉注意力的调制下对目标语音感知有很大贡献。本研究进一步探讨了基于RMS电平的语音分割在竞争性说话者听觉场景中对持续和切换注意力的听觉注意力解码(AAD)性能的影响。还开发了源自皮层活动的客观生物标志物来索引动态听觉注意力状态。在当前研究中,要求受试者在两个竞争性说话者流之间集中或切换注意力。分析了从一个说话者流切换到另一个说话者流之前和之后,对高RMS电平和低RMS电平语音段的神经反应的线性时间响应函数(TRF)。此外,将统一TRF解码模型解码的AAD性能与基于语音RMS电平的分段解码模型在听觉注意力状态动态变化时的性能进行了比较。结果表明,典型TRF分量在大约100毫秒时间滞后的权重对听觉注意力的切换很敏感。与统一AAD模型相比,分段AAD模型在广泛的信号掩蔽比(SMR)下,在持续和切换听觉注意力调制下均提高了注意力解码性能。在竞争性说话者场景中,TRF权重和AAD准确性可作为检测听觉注意力变化的有效指标。此外,在广泛的SMR范围内(即本研究中从6到 -6 dB),分段AAD模型即使在决策窗口长度较短时也表现出强大的解码性能,这表明这种基于语音RMS电平的模型有潜力在现实听觉场景中解码动态注意力状态。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2e6/8866945/ca75e4d9c328/fnins-15-760611-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2e6/8866945/10254dac109c/fnins-15-760611-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2e6/8866945/b091475c8fe2/fnins-15-760611-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2e6/8866945/326a6c200724/fnins-15-760611-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2e6/8866945/ca75e4d9c328/fnins-15-760611-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2e6/8866945/10254dac109c/fnins-15-760611-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2e6/8866945/b091475c8fe2/fnins-15-760611-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2e6/8866945/326a6c200724/fnins-15-760611-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2e6/8866945/ca75e4d9c328/fnins-15-760611-g004.jpg

相似文献

1
A Speech-Level-Based Segmented Model to Decode the Dynamic Auditory Attention States in the Competing Speaker Scenes.一种基于语音水平的分段模型,用于解码竞争说话者场景中的动态听觉注意状态。
Front Neurosci. 2022 Feb 10;15:760611. doi: 10.3389/fnins.2021.760611. eCollection 2021.
2
EEG-based auditory attention decoding using speech-level-based segmented computational models.基于脑电的听觉注意解码,使用基于语音分段的计算模型。
J Neural Eng. 2021 May 25;18(4). doi: 10.1088/1741-2552/abfeba.
3
Robust EEG-Based Decoding of Auditory Attention With High-RMS-Level Speech Segments in Noisy Conditions.基于稳健脑电图的嘈杂环境下高均方根电平语音片段听觉注意力解码
Front Hum Neurosci. 2020 Oct 7;14:557534. doi: 10.3389/fnhum.2020.557534. eCollection 2020.
4
Impact of Different Acoustic Components on EEG-Based Auditory Attention Decoding in Noisy and Reverberant Conditions.不同声成分对噪声和混响环境下基于 EEG 的听觉注意解码的影响。
IEEE Trans Neural Syst Rehabil Eng. 2019 Apr;27(4):652-663. doi: 10.1109/TNSRE.2019.2903404. Epub 2019 Mar 7.
5
Decoding the Attended Speaker From EEG Using Adaptive Evaluation Intervals Captures Fluctuations in Attentional Listening.使用自适应评估间隔从脑电图中解码被关注的说话者可捕捉注意力倾听中的波动。
Front Neurosci. 2020 Jun 16;14:603. doi: 10.3389/fnins.2020.00603. eCollection 2020.
6
Congruent audiovisual speech enhances auditory attention decoding with EEG.视听语音一致增强了 EEG 对听觉注意力的解码。
J Neural Eng. 2019 Nov 6;16(6):066033. doi: 10.1088/1741-2552/ab4340.
7
Robust decoding of the speech envelope from EEG recordings through deep neural networks.通过深度神经网络从 EEG 记录中稳健地解码语音包络。
J Neural Eng. 2022 Jul 6;19(4). doi: 10.1088/1741-2552/ac7976.
8
Neural tracking to go: auditory attention decoding and saliency detection with mobile EEG.神经追踪:使用移动 EEG 进行听觉注意力解码和显著性检测。
J Neural Eng. 2022 Jan 6;18(6). doi: 10.1088/1741-2552/ac42b5.
9
EEG-based auditory attention detection: boundary conditions for background noise and speaker positions.基于脑电图的听觉注意力检测:背景噪声和说话人位置的边界条件。
J Neural Eng. 2018 Dec;15(6):066017. doi: 10.1088/1741-2552/aae0a6. Epub 2018 Sep 12.
10
EEG-based auditory attention decoding with audiovisual speech for hearing-impaired listeners.基于脑电图的听觉注意力解码结合视听语音用于听力受损听众
Cereb Cortex. 2023 Nov 4;33(22):10972-10983. doi: 10.1093/cercor/bhad325.

引用本文的文献

1
Improving auditory attention decoding by classifying intracranial responses to glimpsed and masked acoustic events.通过对瞥见和掩蔽声学事件的颅内反应进行分类来改善听觉注意力解码。
Imaging Neurosci (Camb). 2024;2. doi: 10.1162/imag_a_00148. Epub 2024 Apr 26.

本文引用的文献

1
Dynamic selective auditory attention detection using RNN and reinforcement learning.基于 RNN 和强化学习的动态选择性听觉注意力检测。
Sci Rep. 2021 Jul 29;11(1):15497. doi: 10.1038/s41598-021-94876-0.
2
EEG-based auditory attention decoding using speech-level-based segmented computational models.基于脑电的听觉注意解码,使用基于语音分段的计算模型。
J Neural Eng. 2021 May 25;18(4). doi: 10.1088/1741-2552/abfeba.
3
Neural tracking of the fundamental frequency of the voice: The effect of voice characteristics.神经追踪基频:嗓音特征的影响。
Eur J Neurosci. 2021 Jun;53(11):3640-3653. doi: 10.1111/ejn.15229. Epub 2021 Apr 27.
4
Robust EEG-Based Decoding of Auditory Attention With High-RMS-Level Speech Segments in Noisy Conditions.基于稳健脑电图的嘈杂环境下高均方根电平语音片段听觉注意力解码
Front Hum Neurosci. 2020 Oct 7;14:557534. doi: 10.3389/fnhum.2020.557534. eCollection 2020.
5
Fast EEG-Based Decoding Of The Directional Focus Of Auditory Attention Using Common Spatial Patterns.基于快速脑电图的听觉注意方向的定向空间模式解码。
IEEE Trans Biomed Eng. 2021 May;68(5):1557-1568. doi: 10.1109/TBME.2020.3033446. Epub 2021 Apr 21.
6
Waves of Change: Brain Sensitivity to Differential, not Absolute, Stimulus Intensity is Conserved Across Humans and Rats.变化的浪潮:人类和大鼠的大脑对差异而非绝对刺激强度的敏感性是一致的。
Cereb Cortex. 2021 Jan 5;31(2):949-960. doi: 10.1093/cercor/bhaa267.
7
EEG correlates of spatial shifts of attention in a dynamic multi-talker speech perception scenario in younger and older adults.脑电图对年轻人和老年人在动态多说话人语音感知场景中注意力空间转移的相关性研究。
Hear Res. 2020 Dec;398:108077. doi: 10.1016/j.heares.2020.108077. Epub 2020 Sep 20.
8
Dynamic estimation of auditory temporal response functions via state-space models with Gaussian mixture process noise.基于高斯混合过程噪声的状态空间模型对听觉时频响应函数的动态估计。
PLoS Comput Biol. 2020 Aug 19;16(8):e1008172. doi: 10.1371/journal.pcbi.1008172. eCollection 2020 Aug.
9
Linear versus deep learning methods for noisy speech separation for EEG-informed attention decoding.用于基于脑电图的注意力解码的噪声语音分离的线性与深度学习方法
J Neural Eng. 2020 Aug 19;17(4):046039. doi: 10.1088/1741-2552/aba6f8.
10
Selective auditory attention detection based on effective connectivity by single-trial EEG.基于单试 EEG 的有效连通性的选择性听觉注意力检测。
J Neural Eng. 2020 Apr 17;17(2):026021. doi: 10.1088/1741-2552/ab7c8d.