Du Yuxiao, Ding Han, Wu Min, Chen Feng, Cai Ziman
School of Automation, Guangdong University of Technology, Guangzhou 510006, China.
School of Automation, China University of Geosciences, Wuhan 430074, China.
Brain Sci. 2024 Mar 30;14(4):344. doi: 10.3390/brainsci14040344.
Emotion recognition using the electroencephalogram (EEG) has garnered significant attention within the realm of human-computer interaction due to the wealth of genuine emotional data stored in EEG signals. However, traditional emotion recognition methods are deficient in mining the connection between multi-domain features and fitting their advantages. In this paper, we propose a novel capsule Transformer network based on a multi-domain feature for EEG-based emotion recognition, referred to as MES-CTNet. The model's core consists of a multichannel capsule neural network(CapsNet) embedded with ECA (Efficient Channel Attention) and SE (Squeeze and Excitation) blocks and a Transformer-based temporal coding layer. Firstly, a multi-domain feature map is constructed by combining the space-frequency-time characteristics of the multi-domain features as inputs to the model. Then, the local emotion features are extracted from the multi-domain feature maps by the improved CapsNet. Finally, the Transformer-based temporal coding layer is utilized to globally perceive the emotion feature information of the continuous time slices to obtain a final emotion state. The paper fully experimented on two standard datasets with different emotion labels, the DEAP and SEED datasets. On the DEAP dataset, MES-CTNet achieved an average accuracy of 98.31% in the valence dimension and 98.28% in the arousal dimension; it achieved 94.91% for the cross-session task on the SEED dataset, demonstrating superior performance compared to traditional EEG emotion recognition methods. The MES-CTNet method, utilizing a multi-domain feature map as proposed herein, offers a broader observation perspective for EEG-based emotion recognition. It significantly enhances the classification recognition rate, thereby holding considerable theoretical and practical value in the EEG emotion recognition domain.
由于脑电图(EEG)信号中存储着丰富的真实情感数据,利用脑电图进行情感识别在人机交互领域引起了广泛关注。然而,传统的情感识别方法在挖掘多域特征之间的联系并融合其优势方面存在不足。在本文中,我们提出了一种基于多域特征的新型胶囊Transformer网络用于基于脑电图的情感识别,称为MES-CTNet。该模型的核心由嵌入了ECA(高效通道注意力)和SE(挤压与激励)模块的多通道胶囊神经网络(CapsNet)以及基于Transformer的时间编码层组成。首先,通过将多域特征的空间-频率-时间特征相结合构建多域特征图,作为模型的输入。然后,通过改进的CapsNet从多域特征图中提取局部情感特征。最后,利用基于Transformer的时间编码层全局感知连续时间切片的情感特征信息,以获得最终的情感状态。本文在两个具有不同情感标签的标准数据集DEAP和SEED数据集上进行了充分实验。在DEAP数据集上,MES-CTNet在效价维度上的平均准确率达到98.31%,在唤醒维度上达到98.28%;在SEED数据集的跨会话任务中达到了94.91%,与传统的基于脑电图的情感识别方法相比表现出卓越的性能。本文提出的MES-CTNet方法利用多域特征图,为基于脑电图的情感识别提供了更广阔的观察视角。它显著提高了分类识别率,因此在脑电图情感识别领域具有相当大的理论和实用价值。