• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

脑控增强听觉:在多说话人环境中对空间移动对话的增强

Brain-Controlled Augmented Hearing for Spatially Moving Conversations in Multi-Talker Environments.

机构信息

Department of Electrical Engineering, Columbia University, New York, NY, 10027, USA.

Mortimer B. Zuckerman Mind Brain Behavior Institute, New York, NY, 10027, USA.

出版信息

Adv Sci (Weinh). 2024 Nov;11(41):e2401379. doi: 10.1002/advs.202401379. Epub 2024 Sep 9.

DOI:10.1002/advs.202401379
PMID:39248654
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11538705/
Abstract

Focusing on a specific conversation amidst multiple interfering talkers is challenging, especially for those with hearing loss. Brain-controlled assistive hearing devices aim to alleviate this problem by enhancing the attended speech based on the listener's neural signals using auditory attention decoding (AAD). Departing from conventional AAD studies that relied on oversimplified scenarios with stationary talkers, a realistic AAD task that involves multiple talkers taking turns as they continuously move in space in background noise is presented. Invasive electroencephalography (iEEG) data are collected from three neurosurgical patients as they focused on one of the two moving conversations. An enhanced brain-controlled assistive hearing system that combines AAD and a binaural speaker-independent speech separation model is presented. The separation model unmixes talkers while preserving their spatial location and provides talker trajectories to the neural decoder to improve AAD accuracy. Subjective and objective evaluations show that the proposed system enhances speech intelligibility and facilitates conversation tracking while maintaining spatial cues and voice quality in challenging acoustic environments. This research demonstrates the potential of this approach in real-world scenarios and marks a significant step toward developing assistive hearing technologies that adapt to the intricate dynamics of everyday auditory experiences.

摘要

专注于多个干扰说话者中的特定对话对于有听力损失的人来说具有挑战性。脑控辅助听力设备旨在通过使用听觉注意力解码 (AAD) 根据听众的神经信号增强注意力集中的语音来缓解这个问题。与依赖于固定说话者的简化场景的传统 AAD 研究不同,提出了一个涉及多个说话者在背景噪声中连续在空间中移动并轮流说话的现实 AAD 任务。从三位神经外科患者中采集了侵入性脑电图 (iEEG) 数据,他们专注于两个移动对话中的一个。提出了一种结合 AAD 和双耳说话人独立语音分离模型的增强型脑控辅助听力系统。分离模型在保留说话人空间位置的同时对说话人进行解混,并向神经解码器提供说话人轨迹,以提高 AAD 准确性。主观和客观评估表明,该系统在具有挑战性的声学环境中提高了语音可懂度并促进了对话跟踪,同时保持了空间线索和语音质量。这项研究展示了这种方法在现实场景中的潜力,并朝着开发适应日常听觉体验复杂动态的辅助听力技术迈出了重要一步。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f0ba/11538705/6e15f99fba05/ADVS-11-2401379-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f0ba/11538705/76e3619debc1/ADVS-11-2401379-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f0ba/11538705/2c7c3d185451/ADVS-11-2401379-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f0ba/11538705/aa308119dbc6/ADVS-11-2401379-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f0ba/11538705/b5c811f189d6/ADVS-11-2401379-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f0ba/11538705/7a31f0a04ea2/ADVS-11-2401379-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f0ba/11538705/6e15f99fba05/ADVS-11-2401379-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f0ba/11538705/76e3619debc1/ADVS-11-2401379-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f0ba/11538705/2c7c3d185451/ADVS-11-2401379-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f0ba/11538705/aa308119dbc6/ADVS-11-2401379-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f0ba/11538705/b5c811f189d6/ADVS-11-2401379-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f0ba/11538705/7a31f0a04ea2/ADVS-11-2401379-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f0ba/11538705/6e15f99fba05/ADVS-11-2401379-g002.jpg

相似文献

1
Brain-Controlled Augmented Hearing for Spatially Moving Conversations in Multi-Talker Environments.脑控增强听觉:在多说话人环境中对空间移动对话的增强
Adv Sci (Weinh). 2024 Nov;11(41):e2401379. doi: 10.1002/advs.202401379. Epub 2024 Sep 9.
2
Speaker separation in realistic noise environments with applications to a cognitively-controlled hearing aid.在现实噪声环境中的说话人分离及其在认知控制助听器中的应用。
Neural Netw. 2021 Aug;140:136-147. doi: 10.1016/j.neunet.2021.02.020. Epub 2021 Mar 4.
3
Congruent audiovisual speech enhances auditory attention decoding with EEG.视听语音一致增强了 EEG 对听觉注意力的解码。
J Neural Eng. 2019 Nov 6;16(6):066033. doi: 10.1088/1741-2552/ab4340.
4
Neural decoding of attentional selection in multi-speaker environments without access to clean sources.多说话人环境中无法访问干净源时的注意力选择的神经解码。
J Neural Eng. 2017 Oct;14(5):056001. doi: 10.1088/1741-2552/aa7ab4. Epub 2017 Aug 4.
5
Deep learning-based auditory attention decoding in listeners with hearing impairment.基于深度学习的听力受损者听觉注意力解码
J Neural Eng. 2024 May 22;21(3). doi: 10.1088/1741-2552/ad49d7.
6
EEG-based auditory attention detection: boundary conditions for background noise and speaker positions.基于脑电图的听觉注意力检测:背景噪声和说话人位置的边界条件。
J Neural Eng. 2018 Dec;15(6):066017. doi: 10.1088/1741-2552/aae0a6. Epub 2018 Sep 12.
7
Brain-informed speech separation (BISS) for enhancement of target speaker in multitalker speech perception.基于脑的语音分离(BISS)增强多说话人语音感知中的目标说话人。
Neuroimage. 2020 Dec;223:117282. doi: 10.1016/j.neuroimage.2020.117282. Epub 2020 Aug 20.
8
Noise-robust cortical tracking of attended speech in real-world acoustic scenes.在真实声学场景中对注意到的语音进行抗噪皮层追踪。
Neuroimage. 2017 Aug 1;156:435-444. doi: 10.1016/j.neuroimage.2017.04.026. Epub 2017 Apr 13.
9
Neural decoding of attentional selection in multi-speaker environments without access to separated sources.在无法获取分离声源的多说话者环境中对注意力选择进行神经解码。
Annu Int Conf IEEE Eng Med Biol Soc. 2017 Jul;2017:1644-1647. doi: 10.1109/EMBC.2017.8037155.
10
Improving auditory attention decoding by classifying intracranial responses to glimpsed and masked acoustic events.通过对瞥见和掩蔽声学事件的颅内反应进行分类来改善听觉注意力解码。
Imaging Neurosci (Camb). 2024;2. doi: 10.1162/imag_a_00148. Epub 2024 Apr 26.

引用本文的文献

1
A Brain-Computer Interface for Improving Auditory Attention in Multi-Talker Environments.一种用于在多说话者环境中改善听觉注意力的脑机接口。
bioRxiv. 2025 Mar 13:2025.03.13.641661. doi: 10.1101/2025.03.13.641661.

本文引用的文献

1
A high-performance neuroprosthesis for speech decoding and avatar control.一种用于语音解码和化身控制的高性能神经假体。
Nature. 2023 Aug;620(7976):1037-1046. doi: 10.1038/s41586-023-06443-4. Epub 2023 Aug 23.
2
A high-performance speech neuroprosthesis.高性能言语神经假体
Nature. 2023 Aug;620(7976):1031-1036. doi: 10.1038/s41586-023-06377-x. Epub 2023 Aug 23.
3
ONLINE BINAURAL SPEECH SEPARATION OF MOVING SPEAKERS WITH A WAVESPLIT NETWORK.基于波分裂网络的移动扬声器在线双耳语音分离
Proc IEEE Int Conf Acoust Speech Signal Process. 2023 Jun;2023. doi: 10.1109/icassp49357.2023.10095695. Epub 2023 May 5.
4
Generalizable spelling using a speech neuroprosthesis in an individual with severe limb and vocal paralysis.个体严重的肢体和言语瘫痪中使用言语神经假体实现可泛化的拼写
Nat Commun. 2022 Nov 8;13(1):6510. doi: 10.1038/s41467-022-33611-3.
5
"Unattended, distracting or irrelevant": Theoretical implications of terminological choices in auditory selective attention research.“无人照管、分散注意力或不相关”:听觉选择性注意研究中术语选择的理论影响
Cognition. 2023 Feb;231:105313. doi: 10.1016/j.cognition.2022.105313. Epub 2022 Nov 4.
6
Neural dynamics of phoneme sequences reveal position-invariant code for content and order.音素序列的神经动力学揭示了内容和顺序的位置不变代码。
Nat Commun. 2022 Nov 3;13(1):6606. doi: 10.1038/s41467-022-34326-1.
7
Interaction of bottom-up and top-down neural mechanisms in spatial multi-talker speech perception.上下行神经机制在空间多说话人语音感知中的相互作用。
Curr Biol. 2022 Sep 26;32(18):3971-3986.e4. doi: 10.1016/j.cub.2022.07.047. Epub 2022 Aug 15.
8
EEG alpha and pupil diameter reflect endogenous auditory attention switching and listening effort.脑电图阿尔法和瞳孔直径反映了内源性听觉注意的转换和聆听努力。
Eur J Neurosci. 2022 Mar;55(5):1262-1277. doi: 10.1111/ejn.15616. Epub 2022 Feb 16.
9
Neuroprosthesis for Decoding Speech in a Paralyzed Person with Anarthria.神经假体用于解码无言语症瘫痪患者的言语。
N Engl J Med. 2021 Jul 15;385(3):217-227. doi: 10.1056/NEJMoa2027540.
10
Multi-microphone Complex Spectral Mapping for Utterance-wise and Continuous Speech Separation.用于逐话语和连续语音分离的多麦克风复谱映射
IEEE/ACM Trans Audio Speech Lang Process. 2021;29:2001-2014. doi: 10.1109/taslp.2021.3083405. Epub 2021 May 26.