M V Haresh, K Kannadasan, B Shameedha Begum
Department of Computer Science and Engineering, National Institute of Technology Tiruchirappalli, Tamilnadu, 620015, India.
Behav Brain Res. 2025 Sep 13;493:115652. doi: 10.1016/j.bbr.2025.115652. Epub 2025 Jun 6.
Imagined speech has emerged as a promising paradigm for intuitive control of brain-computer interface (BCI)-based communication systems, providing a means of communication for individuals with severe brain disabilities. In this work, a non-invasive electroencephalogram (EEG)-based automated imagined speech recognition model was proposed to assist communication to convey the individual's intentions or commands. The proposed approach uses Common Spatial Patterns (CSP) and Temporal Patterns (TP) for feature extraction, followed by feature fusion to capture both spatial and temporal dynamics in EEG signals. This fusion of the CSP and TP domains enhances the discriminative power of the extracted features, leading to improved classification accuracy.
An EEG data set was collected from 15 subjects while performing an imagined speech task with a set of ten words that are more suitable for paralyzed patients. The EEG signals were preprocessed and a set of statistical characteristics was extracted from the fused CSP and TP domains. Spectral analysis of the signals was performed with respect to ten imagined words to identify the underlying patterns in EEG. Machine learning models, including Linear Discriminant Analysis (LDA), Random Forest (RF), Support Vector Machine (SVM), and Logistic Regression (LR), were employed for pairwise and multiclass classification.
The proposed model achieved average classification accuracies of 83.83% ± 5.94 and 64.58% ± 10.43 and maximum accuracies of 97.78% and 79.22% for pairwise and multiclass classification, respectively. These results demonstrate the effectiveness of the CSP-TP feature fusion approach, outperforming existing state-of-the-art methods in imagined speech recognition.
The findings suggest that EEG-based automatic imagined speech recognition (AISR) using CSP and TP techniques has significant potential for use in BCI-based assistive technologies, offering a more natural and intuitive means of communication for individuals with severe communication limitations.
想象性言语已成为基于脑机接口(BCI)的通信系统直观控制的一种有前景的范式,为严重脑功能障碍个体提供了一种交流方式。在这项工作中,提出了一种基于非侵入性脑电图(EEG)的自动想象性言语识别模型,以辅助交流来传达个体的意图或指令。所提出的方法使用共同空间模式(CSP)和时间模式(TP)进行特征提取,随后进行特征融合以捕捉EEG信号中的空间和时间动态。CSP和TP域的这种融合增强了提取特征的判别力,从而提高了分类准确率。
从15名受试者收集EEG数据集,同时他们执行一项想象性言语任务,使用一组更适合瘫痪患者的十个单词。对EEG信号进行预处理,并从融合的CSP和TP域中提取一组统计特征。针对十个想象单词对信号进行频谱分析,以识别EEG中的潜在模式。使用包括线性判别分析(LDA)、随机森林(RF)、支持向量机(SVM)和逻辑回归(LR)在内的机器学习模型进行成对和多类分类。
所提出的模型在成对和多类分类中分别实现了83.83%±5.94和64.58%±10.43的平均分类准确率,以及97.78%和79.22%的最高准确率。这些结果证明了CSP-TP特征融合方法的有效性,在想象性言语识别方面优于现有的最先进方法。
研究结果表明,使用CSP和TP技术的基于EEG的自动想象性言语识别(AISR)在基于BCI的辅助技术中有很大的应用潜力,为严重沟通受限的个体提供了一种更自然、直观的交流方式。