• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于稳健听觉脑电信号解码的基于Transformer的对比表征学习

Contrastive representation learning with transformers for robust auditory EEG decoding.

作者信息

Bollens Lies, Accou Bernd, Van Hamme Hugo, Francart Tom

机构信息

Dept. Neurosciences, ExpORL, KU Leuven, Leuven, Belgium.

Dept. of Electrical engineering (ESAT), PSI, KU Leuven, Leuven, Belgium.

出版信息

Sci Rep. 2025 Aug 6;15(1):28744. doi: 10.1038/s41598-025-13646-4.

DOI:10.1038/s41598-025-13646-4
PMID:40770040
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12329020/
Abstract

Decoding of continuous speech from electroencephalography (EEG) presents a promising avenue for understanding neural mechanisms of auditory processing and developing applications in hearing diagnostics. Recent advances in deep learning have improved decoding accuracy. However, challenges remain due to the low signal-to-noise ratio of the recorded brain signals. This study explores the application of contrastive learning, a self-supervised learning technique, to learn robust latent representations of EEG signals. We introduce a novel model architecture that leverages contrastive learning and transformer networks to capture relationships between auditory stimuli and EEG responses. Our model is evaluated on two tasks from the ICASSP 2023 Auditory EEG Decoding Challenge: a binary stimulus classification task (match-mismatch) and stimulus envelope decoding. We achieve state-of-the-art performance on both tasks, significantly outperforming previous winners with 87% accuracy in match-mismatch classification and a 0.176 Pearson correlation in envelope regression. Furthermore, we investigate the impact of model architecture, training set size, and finetuning on decoding performance, providing insights into the factors influencing model generalizability and accuracy. Our findings underscore the potential of contrastive learning for advancing the field of auditory EEG decoding and its potential applications in clinical settings.

摘要

从脑电图(EEG)中解码连续语音为理解听觉处理的神经机制以及开发听力诊断应用提供了一条很有前景的途径。深度学习的最新进展提高了解码准确率。然而,由于记录的脑信号信噪比低,挑战依然存在。本研究探索了对比学习(一种自监督学习技术)在学习EEG信号稳健潜在表征方面的应用。我们引入了一种新颖的模型架构,该架构利用对比学习和Transformer网络来捕捉听觉刺激与EEG反应之间的关系。我们的模型在ICASSP 2023听觉EEG解码挑战赛的两项任务上进行了评估:一个二元刺激分类任务(匹配-不匹配)和刺激包络解码。我们在这两项任务上均取得了领先的性能,在匹配-不匹配分类中以87%的准确率显著超过了之前的获胜者,在包络回归中皮尔逊相关系数达到0.176。此外,我们研究了模型架构、训练集大小和微调对解码性能的影响,为影响模型泛化性和准确性的因素提供了见解。我们的研究结果强调了对比学习在推动听觉EEG解码领域及其在临床环境中的潜在应用方面的潜力。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d43/12329020/161fc46c4812/41598_2025_13646_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d43/12329020/bae85a2133de/41598_2025_13646_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d43/12329020/29bb733b0d73/41598_2025_13646_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d43/12329020/5793603f8662/41598_2025_13646_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d43/12329020/36d22af8655a/41598_2025_13646_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d43/12329020/be55939e0412/41598_2025_13646_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d43/12329020/f06ca15a840f/41598_2025_13646_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d43/12329020/fa2c569900c0/41598_2025_13646_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d43/12329020/161fc46c4812/41598_2025_13646_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d43/12329020/bae85a2133de/41598_2025_13646_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d43/12329020/29bb733b0d73/41598_2025_13646_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d43/12329020/5793603f8662/41598_2025_13646_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d43/12329020/36d22af8655a/41598_2025_13646_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d43/12329020/be55939e0412/41598_2025_13646_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d43/12329020/f06ca15a840f/41598_2025_13646_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d43/12329020/fa2c569900c0/41598_2025_13646_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9d43/12329020/161fc46c4812/41598_2025_13646_Fig8_HTML.jpg

相似文献

1
Contrastive representation learning with transformers for robust auditory EEG decoding.用于稳健听觉脑电信号解码的基于Transformer的对比表征学习
Sci Rep. 2025 Aug 6;15(1):28744. doi: 10.1038/s41598-025-13646-4.
2
Exploring the Potential of Electroencephalography Signal-Based Image Generation Using Diffusion Models: Integrative Framework Combining Mixed Methods and Multimodal Analysis.利用扩散模型探索基于脑电图信号的图像生成潜力:结合混合方法和多模态分析的综合框架
JMIR Med Inform. 2025 Jun 25;13:e72027. doi: 10.2196/72027.
3
Improving auditory attention decoding in noisy environments for listeners with hearing impairment through contrastive learning.通过对比学习改善听力受损者在嘈杂环境中的听觉注意力解码。
J Neural Eng. 2025 Jun 18;22(3). doi: 10.1088/1741-2552/ade28a.
4
AADNet: An End-to-End Deep Learning Model for Auditory Attention Decoding.AADNet:一种用于听觉注意力解码的端到端深度学习模型。
IEEE Trans Neural Syst Rehabil Eng. 2025;33:2695-2706. doi: 10.1109/TNSRE.2025.3587637.
5
Cortical temporal mismatch compensation in bimodal cochlear implant users: Selective attention decoding and pupillometry study.双模人工耳蜗使用者的皮质时间失配补偿:选择性注意解码与瞳孔测量研究。
Hear Res. 2025 Aug;464:109306. doi: 10.1016/j.heares.2025.109306. Epub 2025 May 15.
6
Multi-Class Decoding of Attended Speaker Direction Using Electroencephalogram and Audio Spatial Spectrum.利用脑电图和音频空间频谱对关注的说话者方向进行多类解码。
IEEE Trans Neural Syst Rehabil Eng. 2025;33:2892-2903. doi: 10.1109/TNSRE.2025.3591819.
7
A two-stage EEG zero-shot classification algorithm guided by class reconstruction.一种由类别重构引导的两阶段脑电图零样本分类算法。
J Neural Eng. 2025 Aug 4;22(4). doi: 10.1088/1741-2552/adeaea.
8
Trajectory-Ordered Objectives for Self-Supervised Representation Learning of Temporal Healthcare Data Using Transformers: Model Development and Evaluation Study.使用Transformer进行时间序列医疗数据自监督表示学习的轨迹有序目标:模型开发与评估研究
JMIR Med Inform. 2025 Jun 4;13:e68138. doi: 10.2196/68138.
9
A Contrastive Learning-Enhanced Residual Network for Predicting Epileptic Seizures Using EEG Signals.一种用于利用脑电图信号预测癫痫发作的对比学习增强残差网络。
Int J Neural Syst. 2025 Jul 16:2550050. doi: 10.1142/S0129065725500509.
10
Retrieving and reconstructing conceptually similar images from fMRI with latent diffusion models and a neuro-inspired brain decoding model.使用潜在扩散模型和神经启发式脑解码模型从功能磁共振成像中检索和重建概念上相似的图像。
J Neural Eng. 2024 Jun 28;21(4). doi: 10.1088/1741-2552/ad593c.

本文引用的文献

1
The role of vowel and consonant onsets in neural tracking of natural speech.元音和辅音起音在自然语音神经追踪中的作用。
J Neural Eng. 2024 Jan 11;21(1). doi: 10.1088/1741-2552/ad1784.
2
AttentionViz: A Global View of Transformer Attention.AttentionViz:Transformer注意力机制的全局视图
IEEE Trans Vis Comput Graph. 2024 Jan;30(1):262-272. doi: 10.1109/TVCG.2023.3327163. Epub 2023 Dec 25.
3
Robust neural tracking of linguistic speech representations using a convolutional neural network.使用卷积神经网络实现稳健的语言语音表示的神经跟踪。
J Neural Eng. 2023 Aug 30;20(4). doi: 10.1088/1741-2552/acf1ce.
4
Relating EEG to continuous speech using deep neural networks: a review.利用深度神经网络将 EEG 与连续语音相关联:综述。
J Neural Eng. 2023 Aug 3;20(4). doi: 10.1088/1741-2552/ace73f.
5
EEG Conformer: Convolutional Transformer for EEG Decoding and Visualization.脑电图适配模型:用于脑电图解码与可视化的卷积变换器
IEEE Trans Neural Syst Rehabil Eng. 2023;31:710-719. doi: 10.1109/TNSRE.2022.3230250. Epub 2023 Feb 2.
6
Evidence of a predictive coding hierarchy in the human brain listening to speech.人类大脑在听语音时存在预测编码层级的证据。
Nat Hum Behav. 2023 Mar;7(3):430-441. doi: 10.1038/s41562-022-01516-2. Epub 2023 Mar 2.
7
Decoding of the speech envelope from EEG using the VLAAI deep neural network.使用 VLAAI 深度神经网络对 EEG 进行语音包络解码。
Sci Rep. 2023 Jan 16;13(1):812. doi: 10.1038/s41598-022-27332-2.
8
A hierarchy of linguistic predictions during natural language comprehension.自然语言理解过程中的语言预测层次。
Proc Natl Acad Sci U S A. 2022 Aug 9;119(32):e2201968119. doi: 10.1073/pnas.2201968119. Epub 2022 Aug 3.
9
Robust decoding of the speech envelope from EEG recordings through deep neural networks.通过深度神经网络从 EEG 记录中稳健地解码语音包络。
J Neural Eng. 2022 Jul 6;19(4). doi: 10.1088/1741-2552/ac7976.
10
Deep Correlation Analysis for Audio-EEG Decoding.深度相关分析在音频-脑电图解码中的应用。
IEEE Trans Neural Syst Rehabil Eng. 2021;29:2742-2753. doi: 10.1109/TNSRE.2021.3129790. Epub 2022 Jan 12.