Na In-Seop, Aldrees Asma, Hakeem Abeer, Mohaisen Linda, Umer Muhammad, AlHammadi Dina Abdulaziz, Alsubai Shtwai, Innab Nisreen, Ashraf Imran
Division of Culture Contents, Chonnam National University, Yeosu, Republic of Korea.
Department of Informatics and Computer Systems, College of Computer Science, King Khalid University, Abha, Saudi Arabia.
Front Comput Neurosci. 2024 Dec 11;18:1485121. doi: 10.3389/fncom.2024.1485121. eCollection 2024.
Facial emotion recognition (FER) can serve as a valuable tool for assessing emotional states, which are often linked to mental health. However, mental health encompasses a broad range of factors that go beyond facial expressions. While FER provides insights into certain aspects of emotional well-being, it can be used in conjunction with other assessments to form a more comprehensive understanding of an individual's mental health. This research work proposes a framework for human FER using UNet image segmentation and transfer learning with the EfficientNetB4 model (called FacialNet). The proposed model demonstrates promising results, achieving an accuracy of 90% for six emotion classes (happy, sad, fear, pain, anger, and disgust) and 96.39% for binary classification (happy and sad). The significance of FacialNet is judged by extensive experiments conducted against various machine learning and deep learning models, as well as state-of-the-art previous research works in FER. The significance of FacialNet is further validated using a cross-validation technique, ensuring reliable performance across different data splits. The findings highlight the effectiveness of leveraging UNet image segmentation and EfficientNetB4 transfer learning for accurate and efficient human facial emotion recognition, offering promising avenues for real-world applications in emotion-aware systems and effective computing platforms. Experimental findings reveal that the proposed approach performs substantially better than existing works with an improved accuracy of 96.39% compared to existing 94.26%.
面部表情识别(FER)可作为评估情绪状态的重要工具,而情绪状态往往与心理健康相关。然而,心理健康包含的因素范围广泛,不仅仅局限于面部表情。虽然FER能为情绪健康的某些方面提供见解,但它可与其他评估方法结合使用,以更全面地了解个体的心理健康状况。这项研究工作提出了一个使用UNet图像分割和基于EfficientNetB4模型的迁移学习(称为FacialNet)来进行人类FER的框架。所提出的模型显示出了有前景的结果,对于六种情绪类别(高兴、悲伤、恐惧、痛苦、愤怒和厌恶)的识别准确率达到了90%,对于二元分类(高兴和悲伤)的准确率达到了96.39%。通过与各种机器学习和深度学习模型以及FER领域先前的前沿研究工作进行广泛实验,来判断FacialNet的重要性。使用交叉验证技术进一步验证了FacialNet的重要性,确保在不同数据划分上都有可靠的性能。研究结果突出了利用UNet图像分割和EfficientNetB4迁移学习进行准确高效的人类面部表情识别的有效性,为情感感知系统和有效计算平台在现实世界中的应用提供了有前景的途径。实验结果表明,与现有准确率94.26%相比,所提出的方法准确率提高到96.39%,比现有工作表现得显著更好。