• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

在无法获取分离声源的多说话者环境中对注意力选择进行神经解码。

Neural decoding of attentional selection in multi-speaker environments without access to separated sources.

作者信息

O'Sullivan James, Sheth Sameer A, McKhann Guy, Mehta Ashesh D, Mesgarani Nima

出版信息

Annu Int Conf IEEE Eng Med Biol Soc. 2017 Jul;2017:1644-1647. doi: 10.1109/EMBC.2017.8037155.

DOI:10.1109/EMBC.2017.8037155
PMID:29060199
Abstract

People who suffer from hearing impairments can find it difficult to follow a conversation in a multi-speaker environment. Modern hearing aids can suppress background noise; however, there is little that can be done to help a user attend to a single conversation without knowing which speaker is being attended to. Cognitively controlled hearing aids that use auditory attention decoding (AAD) methods are the next step in offering help. A number of challenges exist, including the lack of access to the clean sound sources in the environment with which to compare with the neural signals. We propose a novel framework that combines single-channel speech separation algorithms with AAD. We present an end-to-end system that 1) receives a single audio channel containing a mixture of speakers that is heard by a listener along with the listener's neural signals, 2) automatically separates the individual speakers in the mixture, 3) determines the attended speaker, and 4) amplifies the attended speaker's voice to assist the listener. Using invasive electrophysiology recordings, our system is able to decode the attention of a subject and detect switches in attention using only the mixed audio. We also identified the regions of the auditory cortex that contribute to AAD. Our quality assessment of the modified audio demonstrates a significant improvement in both subjective and objective speech quality measures. Our novel framework for AAD bridges the gap between the most recent advancements in speech processing technologies and speech prosthesis research and moves us closer to the development of cognitively controlled hearing aids.

摘要

听力受损的人会发现在多说话者环境中很难跟上对话。现代助听器可以抑制背景噪音;然而,在不知道正在关注哪个说话者的情况下,几乎没有什么办法能帮助用户专注于单一对话。使用听觉注意力解码(AAD)方法的认知控制助听器是提供帮助的下一步。存在许多挑战,包括在环境中无法获取干净的声源来与神经信号进行比较。我们提出了一个将单通道语音分离算法与AAD相结合的新颖框架。我们展示了一个端到端系统,该系统:1)接收包含说话者混合声音的单个音频通道,听众能听到这个通道以及听众的神经信号;2)自动分离混合声音中的各个说话者;3)确定被关注的说话者;4)放大被关注说话者的声音以帮助听众。通过侵入性电生理记录,我们的系统能够仅使用混合音频解码受试者的注意力并检测注意力的切换。我们还确定了对AAD有贡献的听觉皮层区域。我们对修改后音频的质量评估表明,在主观和客观语音质量测量方面都有显著改善。我们新颖的AAD框架弥合了语音处理技术和语音假体研究的最新进展之间的差距,并使我们更接近认知控制助听器的开发。

相似文献

1
Neural decoding of attentional selection in multi-speaker environments without access to separated sources.在无法获取分离声源的多说话者环境中对注意力选择进行神经解码。
Annu Int Conf IEEE Eng Med Biol Soc. 2017 Jul;2017:1644-1647. doi: 10.1109/EMBC.2017.8037155.
2
Neural decoding of attentional selection in multi-speaker environments without access to clean sources.多说话人环境中无法访问干净源时的注意力选择的神经解码。
J Neural Eng. 2017 Oct;14(5):056001. doi: 10.1088/1741-2552/aa7ab4. Epub 2017 Aug 4.
3
Speaker separation in realistic noise environments with applications to a cognitively-controlled hearing aid.在现实噪声环境中的说话人分离及其在认知控制助听器中的应用。
Neural Netw. 2021 Aug;140:136-147. doi: 10.1016/j.neunet.2021.02.020. Epub 2021 Mar 4.
4
Speaker-independent auditory attention decoding without access to clean speech sources.无需访问干净语音源的说话人无关听觉注意力解码。
Sci Adv. 2019 May 15;5(5):eaav6134. doi: 10.1126/sciadv.aav6134. eCollection 2019 May.
5
Congruent audiovisual speech enhances auditory attention decoding with EEG.视听语音一致增强了 EEG 对听觉注意力的解码。
J Neural Eng. 2019 Nov 6;16(6):066033. doi: 10.1088/1741-2552/ab4340.
6
EEG-based auditory attention detection: boundary conditions for background noise and speaker positions.基于脑电图的听觉注意力检测:背景噪声和说话人位置的边界条件。
J Neural Eng. 2018 Dec;15(6):066017. doi: 10.1088/1741-2552/aae0a6. Epub 2018 Sep 12.
7
Improved Decoding of Attentional Selection in Multi-Talker Environments with Self-Supervised Learned Speech Representation.利用自监督学习的语音表征改进多说话者环境中注意力选择的解码
Annu Int Conf IEEE Eng Med Biol Soc. 2023 Jul;2023:1-5. doi: 10.1109/EMBC40787.2023.10340191.
8
Selective cortical representation of attended speaker in multi-talker speech perception.选择性皮层对多说话人语音感知中被注意说话人的代表。
Nature. 2012 May 10;485(7397):233-6. doi: 10.1038/nature11020.
9
EEG-Informed Attended Speaker Extraction From Recorded Speech Mixtures With Application in Neuro-Steered Hearing Prostheses.基于脑电图的录制语音混合信号中目标说话者提取及其在神经导向听力假体中的应用
IEEE Trans Biomed Eng. 2017 May;64(5):1045-1056. doi: 10.1109/TBME.2016.2587382. Epub 2016 Jul 7.
10
Brain-Controlled Augmented Hearing for Spatially Moving Conversations in Multi-Talker Environments.脑控增强听觉:在多说话人环境中对空间移动对话的增强
Adv Sci (Weinh). 2024 Nov;11(41):e2401379. doi: 10.1002/advs.202401379. Epub 2024 Sep 9.

引用本文的文献

1
Cognitive component of auditory attention to natural speech events.对自然言语事件听觉注意的认知成分。
Front Hum Neurosci. 2025 Jan 6;18:1460139. doi: 10.3389/fnhum.2024.1460139. eCollection 2024.
2
Comparison of Two-Talker Attention Decoding from EEG with Nonlinear Neural Networks and Linear Methods.两种从 EEG 中解码双说话人注意力的方法比较:非线性神经网络与线性方法。
Sci Rep. 2019 Aug 8;9(1):11538. doi: 10.1038/s41598-019-47795-0.