Feng Xin, Cong Ping, Dong Lin, Xin Yongxian, Miao Fengbo, Xin Ruihao
School of Science, Jilin Institute of Chemical Technology, Jilin, 130000 People's Republic of China.
College of Information and Control Engineering, Jilin Institute of Chemical Technology, Jilin, 130000 People's Republic of China.
Cogn Neurodyn. 2024 Aug;18(4):1689-1707. doi: 10.1007/s11571-023-10034-4. Epub 2023 Dec 4.
Electroencephalogram (EEG) emotion recognition plays a vital role in affective computing. A limitation of the EEG emotion recognition task is that the features of multiple domains are rarely included in the analysis simultaneously because of the lack of an effective feature organization form. This paper proposes a video-level feature organization method to effectively organize the temporal, frequency and spatial domain features. In addition, a deep neural network, Channel Attention Convolutional Aggregation Network, is designed to explore deeper emotional information from video-level features. The network uses a channel attention mechanism to adaptively captures critical EEG frequency bands. Then the frame-level representation of each time point is obtained by multi-layer convolution. Finally, the frame-level features are aggregated through NeXtVLAD to learn the time-sequence-related features. The method proposed in this paper achieves the best classification performance in SEED and DEAP datasets. The mean accuracy and standard deviation of the SEED dataset are 95.80% and 2.04%. In the DEAP dataset, the average accuracy with the standard deviation of arousal and valence are 98.97% ± 1.13% and 98.98% ± 0.98%, respectively. The experimental results show that our approach based on video-level features is effective for EEG emotion recognition tasks.
脑电图(EEG)情感识别在情感计算中起着至关重要的作用。EEG情感识别任务的一个局限性在于,由于缺乏有效的特征组织形式,多领域特征很少同时包含在分析中。本文提出了一种视频级特征组织方法,以有效地组织时域、频域和空域特征。此外,还设计了一种深度神经网络,即通道注意力卷积聚合网络,以从视频级特征中探索更深层次的情感信息。该网络使用通道注意力机制自适应地捕获关键的EEG频段。然后通过多层卷积获得每个时间点的帧级表示。最后,通过NeXtVLAD聚合帧级特征,以学习与时间序列相关的特征。本文提出的方法在SEED和DEAP数据集上取得了最佳分类性能。SEED数据集的平均准确率和标准差分别为95.80%和2.04%。在DEAP数据集中,唤醒度和效价的平均准确率及标准差分别为98.97%±1.13%和98.98%±0.98%。实验结果表明,我们基于视频级特征的方法对EEG情感识别任务是有效的。