• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

具有情境敏感面部表情的交互式数字人的开发。

Development of an Interactive Digital Human with Context-Sensitive Facial Expressions.

作者信息

Yang Fan, Fang Lei, Suo Rui, Zhang Jing, Whang Mincheol

机构信息

Department of Emotion Engineering, Sangmyung University, Seoul 03016, Republic of Korea.

College of Physical Education and Health Engineering, Hebei University of Engineering, Handan 056038, China.

出版信息

Sensors (Basel). 2025 Aug 18;25(16):5117. doi: 10.3390/s25165117.

DOI:10.3390/s25165117
PMID:40871980
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12390055/
Abstract

With the increasing complexity of human-computer interaction scenarios, conventional digital human facial expression systems show notable limitations in handling multi-emotion co-occurrence, dynamic expression, and semantic responsiveness. This paper proposes a digital human system framework that integrates multimodal emotion recognition and compound facial expression generation. The system establishes a complete pipeline for real-time interaction and compound emotional expression, following a sequence of "speech semantic parsing-multimodal emotion recognition-Action Unit (AU)-level 3D facial expression control." First, a ResNet18-based model is employed for robust emotion classification using the AffectNet dataset. Then, an AU motion curve driving module is constructed on the Unreal Engine platform, where dynamic synthesis of basic emotions is achieved via a state-machine mechanism. Finally, Generative Pre-trained Transformer (GPT) is utilized for semantic analysis, generating structured emotional weight vectors that are mapped to the AU layer to enable language-driven facial responses. Experimental results demonstrate that the proposed system significantly improves facial animation quality, with naturalness increasing from 3.54 to 3.94 and semantic congruence from 3.44 to 3.80. These results validate the system's capability to generate realistic and emotionally coherent expressions in real time. This research provides a complete technical framework and practical foundation for high-fidelity digital humans with affective interaction capabilities.

摘要

随着人机交互场景日益复杂,传统数字人面部表情系统在处理多情绪共现、动态表情和语义响应方面存在显著局限性。本文提出了一种集成多模态情感识别和复合面部表情生成的数字人系统框架。该系统建立了一个用于实时交互和复合情感表达的完整流程,遵循“语音语义解析 - 多模态情感识别 - 动作单元(AU) - 3D 面部表情控制”的顺序。首先,使用基于 ResNet18 的模型,利用 AffectNet 数据集进行稳健的情感分类。然后,在虚幻引擎平台上构建一个 AU 运动曲线驱动模块,通过状态机机制实现基本情绪的动态合成。最后,利用生成式预训练变换器(GPT)进行语义分析,生成结构化的情感权重向量,将其映射到 AU 层以实现语言驱动的面部响应。实验结果表明,所提出的系统显著提高了面部动画质量,自然度从 3.54 提高到 3.94,语义一致性从 3.44 提高到 3.80。这些结果验证了该系统实时生成逼真且情感连贯表情的能力。本研究为具有情感交互能力的高保真数字人提供了完整的技术框架和实践基础。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfcc/12390055/721cbd285661/sensors-25-05117-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfcc/12390055/c921be6c8722/sensors-25-05117-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfcc/12390055/0b0b12fb05ed/sensors-25-05117-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfcc/12390055/d043dee7ca4b/sensors-25-05117-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfcc/12390055/761d343a042f/sensors-25-05117-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfcc/12390055/71fcc6a0f0c6/sensors-25-05117-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfcc/12390055/36e36fea40f1/sensors-25-05117-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfcc/12390055/b84cdd1335a1/sensors-25-05117-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfcc/12390055/37aa01f733cf/sensors-25-05117-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfcc/12390055/721cbd285661/sensors-25-05117-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfcc/12390055/c921be6c8722/sensors-25-05117-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfcc/12390055/0b0b12fb05ed/sensors-25-05117-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfcc/12390055/d043dee7ca4b/sensors-25-05117-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfcc/12390055/761d343a042f/sensors-25-05117-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfcc/12390055/71fcc6a0f0c6/sensors-25-05117-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfcc/12390055/36e36fea40f1/sensors-25-05117-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfcc/12390055/b84cdd1335a1/sensors-25-05117-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfcc/12390055/37aa01f733cf/sensors-25-05117-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dfcc/12390055/721cbd285661/sensors-25-05117-g009.jpg

相似文献

1
Development of an Interactive Digital Human with Context-Sensitive Facial Expressions.具有情境敏感面部表情的交互式数字人的开发。
Sensors (Basel). 2025 Aug 18;25(16):5117. doi: 10.3390/s25165117.
2
Facial Emotion Recognition of 16 Distinct Emotions From Smartphone Videos: Comparative Study of Machine Learning and Human Performance.基于智能手机视频的16种不同情绪的面部表情识别:机器学习与人类表现的对比研究
J Med Internet Res. 2025 Jul 2;27:e68942. doi: 10.2196/68942.
3
Cognitive decline assessment using semantic linguistic content and transformer deep learning architecture.使用语义语言内容和变压器深度学习架构评估认知能力下降。
Int J Lang Commun Disord. 2024 May-Jun;59(3):1110-1127. doi: 10.1111/1460-6984.12973. Epub 2023 Nov 16.
4
Data Mining-Based Model for Computer-Aided Diagnosis of Autism and Gelotophobia: Mixed Methods Deep Learning Approach.基于数据挖掘的自闭症和恐笑症计算机辅助诊断模型:混合方法深度学习途径
JMIR Form Res. 2025 Aug 13;9:e72115. doi: 10.2196/72115.
5
Integrated neural network framework for multi-object detection and recognition using UAV imagery.用于使用无人机图像进行多目标检测与识别的集成神经网络框架。
Front Neurorobot. 2025 Jul 30;19:1643011. doi: 10.3389/fnbot.2025.1643011. eCollection 2025.
6
Action unit based micro-expression recognition framework for driver emotional state detection.基于动作单元的驾驶员情绪状态检测微表情识别框架
Sci Rep. 2025 Jul 30;15(1):27824. doi: 10.1038/s41598-025-12245-7.
7
Facial emotion recognition accuracy in women with symptoms of polycystic ovary syndrome: Reduced fear and disgust perception.患有多囊卵巢综合征症状女性的面部情绪识别准确性:恐惧和厌恶感知能力降低。
Womens Health (Lond). 2025 Jan-Dec;21:17455057251359761. doi: 10.1177/17455057251359761. Epub 2025 Jul 28.
8
Development of a Cohesive Predictive Model for Substance Use Disorder Rehabilitation Using Passive Digital Biomarkers, Psychological Assessments, and Automated Facial Emotion Recognition: Protocol for a Prospective Cohort Study.使用被动数字生物标志物、心理评估和自动面部情绪识别开发物质使用障碍康复的连贯预测模型:一项前瞻性队列研究方案
JMIR Res Protoc. 2025 Jun 27;14:e71374. doi: 10.2196/71374.
9
Feasibility study of emotion mimicry analysis in human-machine interaction.人机交互中情感模仿分析的可行性研究。
Sci Rep. 2025 Jan 31;15(1):3859. doi: 10.1038/s41598-025-87688-z.
10
Three-Dimensional View Relationship-Based Context-Aware Emotion Recognition.
IEEE Trans Neural Netw Learn Syst. 2025 Jul;36(7):13567-13578. doi: 10.1109/TNNLS.2024.3476249.

本文引用的文献

1
Experiments on real-life emotions challenge Ekman's model.对真实情感的实验挑战了埃克曼的模型。
Sci Rep. 2023 Jun 12;13(1):9511. doi: 10.1038/s41598-023-36201-5.
2
Devising novel performance measures for assessing the behavior of multilayer perceptrons trained on regression tasks.设计用于评估基于回归任务训练的多层感知机行为的新型性能指标。
PLoS One. 2023 May 18;18(5):e0285471. doi: 10.1371/journal.pone.0285471. eCollection 2023.
3
GANimation: Anatomically-aware Facial Animation from a Single Image.GANimation:基于单张图像的解剖学感知面部动画
Comput Vis ECCV. 2018 Sep;11214:835-851. doi: 10.1007/978-3-030-01249-6_50. Epub 2018 Oct 6.
4
Exploring expressivity and emotion with artificial voice and speech technologies.利用人工语音和语音技术探索表现力与情感。
Logoped Phoniatr Vocol. 2013 Oct;38(3):115-25. doi: 10.3109/14015439.2013.810303. Epub 2013 Sep 12.
5
FACSGen 2.0 animation software: generating three-dimensional FACS-valid facial expressions for emotion research.FACSGen 2.0 动画软件:为情感研究生成三维 FACS 有效面部表情。
Emotion. 2012 Apr;12(2):351-63. doi: 10.1037/a0026632. Epub 2012 Jan 16.
6
Introducing the Geneva Multimodal expression corpus for experimental research on emotion perception.介绍用于情绪感知实验研究的日内瓦多模态表达语料库。
Emotion. 2012 Oct;12(5):1161-79. doi: 10.1037/a0025827. Epub 2011 Nov 14.
7
A new physical model with multilayer architecture for facial expression animation using dynamic adaptive mesh.一种用于面部表情动画的具有多层架构的新物理模型,采用动态自适应网格。
IEEE Trans Vis Comput Graph. 2004 May-Jun;10(3):339-52. doi: 10.1109/TVCG.2004.1272733.
8
Constants across cultures in the face and emotion.面部与情感方面的跨文化常量。
J Pers Soc Psychol. 1971 Feb;17(2):124-9. doi: 10.1037/h0030377.