IEEE J Biomed Health Inform. 2022 Feb;26(2):589-599. doi: 10.1109/JBHI.2021.3092412. Epub 2022 Feb 4.
With the development of sensor technology and learning algorithms, multimodal emotion recognition has attracted widespread attention. Many existing studies on emotion recognition mainly focused on normal people. Besides, due to hearing loss, deaf people cannot express emotions by words, which may have a greater need for emotion recognition. In this paper, the deep belief network (DBN) was utilized to classify three category emotions through the electroencephalograph (EEG) and facial expressions. Signals from 15 deaf subjects were recorded when they watched the emotional movie clips. Our system uses a 1-s window without overlap to segment the EEG signals in five frequency bands, then the differential entropy (DE) feature is extracted. The DE feature of EEG and facial expression images plays as multimodal input for subject-dependent emotion recognition. To avoid feature redundancy, the top 12 major EEG electrode channels (FP2, FP1, FT7, FPZ, F7, T8, F8, CB2, CB1, FT8, T7, TP8) in the gamma band and 30 facial expression features (the areas around the eyes and eyebrow) which are selected by the largest weight values. The results show that the classification accuracy is 99.92% by feature selection in deaf emotion reignition. Moreover, investigations on brain activities reveal deaf brain activity changes mainly in the beta and gamma bands, and the brain regions that are affected by emotions are mainly distributed in the prefrontal and outer temporal lobes.
随着传感器技术和学习算法的发展,多模态情绪识别引起了广泛关注。许多现有的情绪识别研究主要集中在正常人身上。此外,由于听力损失,聋人无法用言语表达情绪,因此他们可能更需要情绪识别。在本文中,利用深度置信网络(DBN)通过脑电图(EEG)和面部表情对三类情绪进行分类。当 15 名聋人观看情感电影片段时,记录了他们的信号。我们的系统使用无重叠的 1 秒窗口来分割五个频带的 EEG 信号,然后提取差分熵(DE)特征。EEG 和面部表情图像的 DE 特征作为受试者相关情绪识别的多模态输入。为了避免特征冗余,在伽马波段选择了前 12 个主要 EEG 电极通道(FP2、FP1、FT7、FPZ、F7、T8、F8、CB2、CB1、FT8、T7、TP8)和 30 个面部表情特征(眼睛和眉毛周围的区域),这些特征是通过最大权重值选择的。结果表明,特征选择在聋人情绪重燃中的分类准确率为 99.92%。此外,对大脑活动的研究表明,聋人大脑活动的变化主要发生在β和γ波段,受情绪影响的大脑区域主要分布在前额叶和外颞叶。