Liu Dongjun, Cui Jin, Pan Zeyu, Zhang Hangkui, Cao Jianting, Kong Wanzeng
School of Computer Science, Hangzhou Dianzi University, Hangzhou, 310018 Zhejiang China.
Graduate School of Engineering, Saitama Institute of Technology, Saitama, 369-0293 Japan.
Cogn Neurodyn. 2024 Jun;18(3):863-875. doi: 10.1007/s11571-023-09946-y. Epub 2023 Feb 22.
The human brain can effectively perform Facial Expression Recognition (FER) with a few samples by utilizing its cognitive ability. However, unlike the human brain, even the well-trained deep neural network is data-dependent and lacks cognitive ability. To tackle this challenge, this paper proposes a novel framework, Brain Machine Generative Adversarial Networks (BM-GAN), which utilizes the concept of brain's cognitive ability to guide a Convolutional Neural Network to generate LIKE-electroencephalograph (EEG) features. More specifically, we firstly obtain EEG signals triggered from facial emotion images, then we adopt BM-GAN to carry out the mutual generation of image visual features and EEG cognitive features. BM-GAN intends to use the cognitive knowledge learnt from EEG signals to instruct the model to perceive LIKE-EEG features. Thereby, BM-GAN has a superior performance for FER like the human brain. The proposed model consists of VisualNet, EEGNet, and BM-GAN. More specifically, VisualNet can obtain image visual features from facial emotion images and EEGNet can obtain EEG cognitive features from EEG signals. Subsequently, the BM-GAN completes the mutual generation of image visual features and EEG cognitive features. Finally, the predicted LIKE-EEG features of test images are used for FER. After learning, without the participation of the EEG signals, an average classification accuracy of 96.6 % is obtained on Chinese Facial Affective Picture System dataset using LIKE-EEG features for FER. Experiments demonstrate that the proposed method can produce an excellent performance for FER.
人类大脑能够利用其认知能力,通过少量样本有效地进行面部表情识别(FER)。然而,与人类大脑不同的是,即使是训练有素的深度神经网络也依赖数据且缺乏认知能力。为应对这一挑战,本文提出了一种新颖的框架——脑机生成对抗网络(BM-GAN),该框架利用大脑认知能力的概念来引导卷积神经网络生成类脑电图(EEG)特征。具体而言,我们首先获取由面部情感图像触发的EEG信号,然后采用BM-GAN进行图像视觉特征和EEG认知特征的相互生成。BM-GAN旨在利用从EEG信号中学到的认知知识来指导模型感知类EEG特征。因此,BM-GAN在FER方面具有像人类大脑一样的卓越性能。所提出的模型由视觉网络(VisualNet)、脑电图网络(EEGNet)和BM-GAN组成。具体来说,VisualNet可以从面部情感图像中获取图像视觉特征,EEGNet可以从EEG信号中获取EEG认知特征。随后,BM-GAN完成图像视觉特征和EEG认知特征的相互生成。最后,将测试图像的预测类EEG特征用于FER。经过学习,在不使用EEG信号参与的情况下,使用类EEG特征在中文面部表情图片系统数据集上进行FER时,平均分类准确率达到了96.6%。实验表明,所提出的方法在FER方面能够产生优异的性能。