School of Software, South China Normal University, Guangzhou 510641, China.
Comput Intell Neurosci. 2017;2017:2107451. doi: 10.1155/2017/2107451. Epub 2017 Sep 19.
This paper proposes two multimodal fusion methods between brain and peripheral signals for emotion recognition. The input signals are electroencephalogram and facial expression. The stimuli are based on a subset of movie clips that correspond to four specific areas of valance-arousal emotional space (happiness, neutral, sadness, and fear). For facial expression detection, four basic emotion states (happiness, neutral, sadness, and fear) are detected by a neural network classifier. For EEG detection, four basic emotion states and three emotion intensity levels (strong, ordinary, and weak) are detected by two support vector machines (SVM) classifiers, respectively. Emotion recognition is based on two decision-level fusion methods of both EEG and facial expression detections by using a sum rule or a production rule. Twenty healthy subjects attended two experiments. The results show that the accuracies of two multimodal fusion detections are 81.25% and 82.75%, respectively, which are both higher than that of facial expression (74.38%) or EEG detection (66.88%). The combination of facial expressions and EEG information for emotion recognition compensates for their defects as single information sources.
本文提出了两种脑与外周信号的多模态融合方法,用于情感识别。输入信号是脑电图和面部表情。刺激基于对应于效价唤醒情感空间的四个特定区域的电影片段子集(快乐、中性、悲伤和恐惧)。对于面部表情检测,通过神经网络分类器检测到四种基本情绪状态(快乐、中性、悲伤和恐惧)。对于 EEG 检测,通过两个支持向量机 (SVM) 分类器分别检测到四种基本情绪状态和三种情绪强度水平(强、中、弱)。情感识别基于 EEG 和面部表情检测的两种决策级融合方法,使用求和规则或产生式规则。二十名健康受试者参加了两项实验。结果表明,两种多模态融合检测的准确率分别为 81.25%和 82.75%,均高于面部表情(74.38%)或脑电图检测(66.88%)。表情和脑电图信息的组合用于情感识别,弥补了它们作为单一信息源的缺陷。