• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

通过对瞥见和掩蔽声学事件的颅内反应进行分类来改善听觉注意力解码。

Improving auditory attention decoding by classifying intracranial responses to glimpsed and masked acoustic events.

作者信息

Raghavan Vinay S, O'Sullivan James, Herrero Jose, Bickel Stephan, Mehta Ashesh D, Mesgarani Nima

机构信息

Department of Electrical Engineering, Columbia University, New York, NY, United States.

Zuckerman Mind Brain Behavior Institute, Columbia University, New York, NY, United States.

出版信息

Imaging Neurosci (Camb). 2024;2. doi: 10.1162/imag_a_00148. Epub 2024 Apr 26.

DOI:10.1162/imag_a_00148
PMID:39867597
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11759098/
Abstract

Listeners with hearing loss have trouble following a conversation in multitalker environments. While modern hearing aids can generally amplify speech, these devices are unable to tune into a target speaker without first knowing to which speaker a user aims to attend. Brain-controlled hearing aids have been proposed using auditory attention decoding (AAD) methods, but current methods use the same model to compare the speech stimulus and neural response, regardless of the dynamic overlap between talkers which is known to influence neural encoding. Here, we propose a novel framework that directly classifies event-related potentials (ERPs) evoked by glimpsed and masked acoustic events to determine whether the source of the event was attended. We present a system that identifies auditory events using the local maxima in the envelope rate of change, assesses the temporal masking of auditory events relative to competing speakers, and utilizes masking-specific ERP classifiers to determine if the source of the event was attended. Using intracranial electrophysiological recordings, we showed that high gamma ERPs from recording sites in auditory cortex can effectively decode the attention of subjects. This method of AAD provides higher accuracy, shorter switch times, and more stable decoding results compared with traditional correlational methods, permitting the quick and accurate detection of changes in a listener's attentional focus. This framework also holds unique potential for detecting instances of divided attention and inattention. Overall, we extend the scope of AAD algorithms by introducing the first linear, direct-classification method for determining a listener's attentional focus that leverages the latest research in multitalker speech perception. This work represents another step toward informing the development of effective and intuitive brain-controlled hearing assistive devices.

摘要

听力损失患者在多说话者环境中难以跟上对话。虽然现代助听器通常可以放大语音,但这些设备在不知道用户想要关注哪个说话者的情况下,无法锁定目标说话者。有人提出使用听觉注意力解码(AAD)方法的脑控助听器,但目前的方法使用相同的模型来比较语音刺激和神经反应,而忽略了已知会影响神经编码的说话者之间的动态重叠。在这里,我们提出了一个新颖的框架,该框架直接对瞥见和掩蔽声学事件诱发的事件相关电位(ERP)进行分类,以确定事件的来源是否被关注。我们提出了一个系统,该系统使用包络变化率的局部最大值来识别听觉事件,评估听觉事件相对于竞争说话者的时间掩蔽,并利用特定于掩蔽的ERP分类器来确定事件的来源是否被关注。使用颅内电生理记录,我们表明来自听觉皮层记录部位的高伽马ERP可以有效地解码受试者的注意力。与传统的相关方法相比,这种AAD方法提供了更高的准确性、更短的切换时间和更稳定的解码结果,能够快速准确地检测听众注意力焦点的变化。该框架在检测注意力分散和注意力不集中的情况方面也具有独特的潜力。总体而言,我们通过引入第一种线性直接分类方法来扩展AAD算法的范围,该方法用于确定听众的注意力焦点,利用了多说话者语音感知的最新研究成果。这项工作朝着为有效且直观的脑控听力辅助设备的开发提供信息又迈出了一步。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/20b3/11759098/124678996669/nihms-2009028-f0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/20b3/11759098/43d862762e32/nihms-2009028-f0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/20b3/11759098/6a0a134b1042/nihms-2009028-f0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/20b3/11759098/281b06852df7/nihms-2009028-f0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/20b3/11759098/cd3e211b3efa/nihms-2009028-f0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/20b3/11759098/e7b2e4efd11a/nihms-2009028-f0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/20b3/11759098/124678996669/nihms-2009028-f0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/20b3/11759098/43d862762e32/nihms-2009028-f0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/20b3/11759098/6a0a134b1042/nihms-2009028-f0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/20b3/11759098/281b06852df7/nihms-2009028-f0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/20b3/11759098/cd3e211b3efa/nihms-2009028-f0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/20b3/11759098/e7b2e4efd11a/nihms-2009028-f0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/20b3/11759098/124678996669/nihms-2009028-f0006.jpg

相似文献

1
Improving auditory attention decoding by classifying intracranial responses to glimpsed and masked acoustic events.通过对瞥见和掩蔽声学事件的颅内反应进行分类来改善听觉注意力解码。
Imaging Neurosci (Camb). 2024;2. doi: 10.1162/imag_a_00148. Epub 2024 Apr 26.
2
Neural decoding of attentional selection in multi-speaker environments without access to clean sources.多说话人环境中无法访问干净源时的注意力选择的神经解码。
J Neural Eng. 2017 Oct;14(5):056001. doi: 10.1088/1741-2552/aa7ab4. Epub 2017 Aug 4.
3
Neural decoding of attentional selection in multi-speaker environments without access to separated sources.在无法获取分离声源的多说话者环境中对注意力选择进行神经解码。
Annu Int Conf IEEE Eng Med Biol Soc. 2017 Jul;2017:1644-1647. doi: 10.1109/EMBC.2017.8037155.
4
Congruent audiovisual speech enhances auditory attention decoding with EEG.视听语音一致增强了 EEG 对听觉注意力的解码。
J Neural Eng. 2019 Nov 6;16(6):066033. doi: 10.1088/1741-2552/ab4340.
5
Brain-Controlled Augmented Hearing for Spatially Moving Conversations in Multi-Talker Environments.脑控增强听觉:在多说话人环境中对空间移动对话的增强
Adv Sci (Weinh). 2024 Nov;11(41):e2401379. doi: 10.1002/advs.202401379. Epub 2024 Sep 9.
6
EEG-based auditory attention decoding using speech-level-based segmented computational models.基于脑电的听觉注意解码,使用基于语音分段的计算模型。
J Neural Eng. 2021 May 25;18(4). doi: 10.1088/1741-2552/abfeba.
7
EEG-based auditory attention detection: boundary conditions for background noise and speaker positions.基于脑电图的听觉注意力检测:背景噪声和说话人位置的边界条件。
J Neural Eng. 2018 Dec;15(6):066017. doi: 10.1088/1741-2552/aae0a6. Epub 2018 Sep 12.
8
A Speech-Level-Based Segmented Model to Decode the Dynamic Auditory Attention States in the Competing Speaker Scenes.一种基于语音水平的分段模型,用于解码竞争说话者场景中的动态听觉注意状态。
Front Neurosci. 2022 Feb 10;15:760611. doi: 10.3389/fnins.2021.760611. eCollection 2021.
9
Impact of Different Acoustic Components on EEG-Based Auditory Attention Decoding in Noisy and Reverberant Conditions.不同声成分对噪声和混响环境下基于 EEG 的听觉注意解码的影响。
IEEE Trans Neural Syst Rehabil Eng. 2019 Apr;27(4):652-663. doi: 10.1109/TNSRE.2019.2903404. Epub 2019 Mar 7.
10
Neural tracking to go: auditory attention decoding and saliency detection with mobile EEG.神经追踪:使用移动 EEG 进行听觉注意力解码和显著性检测。
J Neural Eng. 2022 Jan 6;18(6). doi: 10.1088/1741-2552/ac42b5.

本文引用的文献

1
Single-neuronal elements of speech production in humans.人类言语产生中单神经元的成分。
Nature. 2024 Feb;626(7999):603-610. doi: 10.1038/s41586-023-06982-w. Epub 2024 Jan 31.
2
What are wedecoding? Unveiling biases in EEG-based decoding of the spatial focus of auditory attention.我们正在解码什么?揭示基于 EEG 的听觉注意力空间焦点解码中的偏差。
J Neural Eng. 2024 Feb 6;21(1). doi: 10.1088/1741-2552/ad2214.
3
Large-scale single-neuron speech sound encoding across the depth of human cortex.人类大脑皮层深度范围内的大规模单神经元语音编码
Nature. 2024 Feb;626(7999):593-602. doi: 10.1038/s41586-023-06839-2. Epub 2023 Dec 13.
4
naplib-python: Neural acoustic data processing and analysis tools in python.naplib-python:Python 中的神经声学数据处理与分析工具。
Softw Impacts. 2023 Sep;17. doi: 10.1016/j.simpa.2023.100541. Epub 2023 Jul 7.
5
A high-performance neuroprosthesis for speech decoding and avatar control.一种用于语音解码和化身控制的高性能神经假体。
Nature. 2023 Aug;620(7976):1037-1046. doi: 10.1038/s41586-023-06443-4. Epub 2023 Aug 23.
6
A high-performance speech neuroprosthesis.高性能言语神经假体
Nature. 2023 Aug;620(7976):1031-1036. doi: 10.1038/s41586-023-06377-x. Epub 2023 Aug 23.
7
Distinct neural encoding of glimpsed and masked speech in multitalker situations.多说话人情况下瞥见和掩蔽语音的神经编码特征不同。
PLoS Biol. 2023 Jun 6;21(6):e3002128. doi: 10.1371/journal.pbio.3002128. eCollection 2023 Jun.
8
Timing in Conversation.对话中的时机把握。
J Cogn. 2023 Apr 5;6(1):20. doi: 10.5334/joc.268. eCollection 2023.
9
Listening to two speakers: Capacity and tradeoffs in neural speech tracking during Selective and Distributed Attention.倾听两位说话者:选择性和分布式注意过程中神经语音追踪的能力与权衡
Neuroimage. 2023 Apr 15;270:119984. doi: 10.1016/j.neuroimage.2023.119984. Epub 2023 Feb 26.
10
Speech onsets and sustained speech contribute differentially to delta and theta speech tracking in auditory cortex.言语起始和持续言语对听觉皮层中 delta 和 theta 言语跟踪的贡献不同。
Cereb Cortex. 2023 May 9;33(10):6273-6281. doi: 10.1093/cercor/bhac502.