Hou Xiaoyuan, Zhao Jing, Zhang Hui
School of Engineering Medicine, Beihang University, Beijing, China.
School of Biological Science and Medical Engineering, Beihang University, Beijing, China.
Front Neurosci. 2022 Oct 26;16:1015752. doi: 10.3389/fnins.2022.1015752. eCollection 2022.
Reconstruction of perceived faces from brain signals is a hot topic in brain decoding and an important application in the field of brain-computer interfaces. Existing methods do not fully consider the multiple facial attributes represented in face images, and their different activity patterns at multiple brain regions are often ignored, which causes the reconstruction performance very poor. In the current study, we propose an algorithmic framework that efficiently combines multiple face-selective brain regions for precise multi-attribute perceived face reconstruction. Our framework consists of three modules: a multi-task deep learning network (MTDLN), which is developed to simultaneously extract the multi-dimensional face features attributed to facial expression, identity and gender from one single face image, a set of linear regressions (LR), which is built to map the relationship between the multi-dimensional face features and the brain signals from multiple brain regions, and a multi-conditional generative adversarial network (mcGAN), which is used to generate the perceived face images constrained by the predicted multi-dimensional face features. We conduct extensive fMRI experiments to evaluate the reconstruction performance of our framework both subjectively and objectively. The results show that, compared with the traditional methods, our proposed framework better characterizes the multi-attribute face features in a face image, better predicts the face features from brain signals, and achieves better reconstruction performance of both seen and unseen face images in both visual effects and quantitative assessment. Moreover, besides the state-of-the-art intra-subject reconstruction performance, our proposed framework can also realize inter-subject face reconstruction to a certain extent.
从脑信号重建感知面孔是脑解码领域的一个热门话题,也是脑机接口领域的一项重要应用。现有方法没有充分考虑面部图像中所呈现的多种面部属性,并且常常忽略它们在多个脑区的不同活动模式,这导致重建性能非常差。在当前研究中,我们提出了一个算法框架,该框架有效地结合多个面部选择性脑区,用于精确的多属性感知面孔重建。我们的框架由三个模块组成:一个多任务深度学习网络(MTDLN),其被开发用于从单张面部图像中同时提取归因于面部表情、身份和性别的多维度面部特征;一组线性回归(LR),其被构建用于映射多维度面部特征与来自多个脑区的脑信号之间的关系;以及一个多条件生成对抗网络(mcGAN),其用于生成受预测的多维度面部特征约束的感知面部图像。我们进行了广泛的功能磁共振成像实验,以主观和客观地评估我们框架的重建性能。结果表明,与传统方法相比,我们提出的框架能更好地表征面部图像中的多属性面部特征,能更好地从脑信号中预测面部特征,并且在视觉效果和定量评估方面对已见和未见面部图像均实现了更好的重建性能。此外,除了具有最先进的受试者内重建性能外,我们提出的框架在一定程度上还能实现受试者间的面孔重建。