• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

通过学习语音协同发音和表情空间实现表情丰富的面部动画合成。

Expressive facial animation synthesis by learning speech coarticulation and expression spaces.

作者信息

Deng Zhigang, Neumann Ulrich, Lewis J P, Kim Tae-Yong, Bulut Murtaza, Narayanan Shrikanth

机构信息

Department of Computer Science, University of Houston, TX 77004, USA.

出版信息

IEEE Trans Vis Comput Graph. 2006 Nov-Dec;12(6):1523-34. doi: 10.1109/TVCG.2006.90.

DOI:10.1109/TVCG.2006.90
PMID:17073374
Abstract

Synthesizing expressive facial animation is a very challenging topic within the graphics community. In this paper, we present an expressive facial animation synthesis system enabled by automated learning from facial motion capture data. Accurate 3D motions of the markers on the face of a human subject are captured while he/she recites a predesigned corpus, with specific spoken and visual expressions. We present a novel motion capture mining technique that "learns" speech coarticulation models for diphones and triphones from the recorded data. A Phoneme-Independent Expression Eigenspace (PIEES) that encloses the dynamic expression signals is constructed by motion signal processing (phoneme-based time-warping and subtraction) and Principal Component Analysis (PCA) reduction. New expressive facial animations are synthesized as follows: First, the learned coarticulation models are concatenated to synthesize neutral visual speech according to novel speech input, then a texture-synthesis-based approach is used to generate a novel dynamic expression signal from the PIEES model, and finally the synthesized expression signal is blended with the synthesized neutral visual speech to create the final expressive facial animation. Our experiments demonstrate that the system can effectively synthesize realistic expressive facial animation.

摘要

合成富有表现力的面部动画是图形学界一个极具挑战性的课题。在本文中,我们提出了一个通过从面部动作捕捉数据中自动学习来实现的富有表现力的面部动画合成系统。当人类受试者背诵预先设计的语料库时,同时带有特定的语音和视觉表情,面部标记的精确3D运动被捕捉下来。我们提出了一种新颖的动作捕捉挖掘技术,该技术从记录的数据中“学习”双音素和三音素的语音协同发音模型。通过运动信号处理(基于音素的时间规整和减法)和主成分分析(PCA)降维,构建了一个包含动态表情信号的音素无关表情特征空间(PIEES)。新的富有表现力的面部动画合成如下:首先,根据新的语音输入连接学习到的协同发音模型来合成中性视觉语音,然后使用基于纹理合成的方法从PIEES模型生成新的动态表情信号,最后将合成的表情信号与合成的中性视觉语音混合以创建最终的富有表现力的面部动画。我们的实验表明,该系统能够有效地合成逼真的富有表现力的面部动画。

相似文献

1
Expressive facial animation synthesis by learning speech coarticulation and expression spaces.通过学习语音协同发音和表情空间实现表情丰富的面部动画合成。
IEEE Trans Vis Comput Graph. 2006 Nov-Dec;12(6):1523-34. doi: 10.1109/TVCG.2006.90.
2
Accurate visible speech synthesis based on concatenating variable length motion capture data.基于拼接可变长度动作捕捉数据的精确可视语音合成。
IEEE Trans Vis Comput Graph. 2006 Mar-Apr;12(2):266-76. doi: 10.1109/TVCG.2006.18.
3
Creating speech-synchronized animation.创建语音同步动画。
IEEE Trans Vis Comput Graph. 2005 May-Jun;11(3):341-52. doi: 10.1109/TVCG.2005.43.
4
Orthogonal-blendshape-based editing system for facial motion capture data.基于正交混合形状的面部动作捕捉数据编辑系统。
IEEE Comput Graph Appl. 2008 Nov-Dec;28(6):76-82. doi: 10.1109/MCG.2008.120.
5
Modeling short-term dynamics and variability for realistic interactive facial animation.为逼真的交互式面部动画建模短期动态和变异性。
IEEE Comput Graph Appl. 2010 Jul-Aug;30(4):51-61. doi: 10.1109/MCG.2010.40.
6
Transferring of speech movements from video to 3D face space.将语音动作从视频转移到3D面部空间。
IEEE Trans Vis Comput Graph. 2007 Jan-Feb;13(1):58-69. doi: 10.1109/TVCG.2007.22.
7
Geometry-driven photorealistic facial expression synthesis.基于几何驱动的逼真面部表情合成。
IEEE Trans Vis Comput Graph. 2006 Jan-Feb;12(1):48-60. doi: 10.1109/TVCG.2006.9.
8
Three-dimensional face recognition in the presence of facial expressions: an annotated deformable model approach.存在面部表情时的三维人脸识别:一种带注释的可变形模型方法。
IEEE Trans Pattern Anal Mach Intell. 2007 Apr;29(4):640-9. doi: 10.1109/TPAMI.2007.1017.
9
The computer synthesis of expressive faces.富有表现力的面部的计算机合成。
Philos Trans R Soc Lond B Biol Sci. 1992 Jan 29;335(1273):87-93. doi: 10.1098/rstb.1992.0011.
10
Analysis of head gesture and prosody patterns for prosody-driven head-gesture animation.用于韵律驱动头部手势动画的头部手势和韵律模式分析。
IEEE Trans Pattern Anal Mach Intell. 2008 Aug;30(8):1330-45. doi: 10.1109/TPAMI.2007.70797.

引用本文的文献

1
Generating human facial animation by aggregation deep network and low-rank active learning with table tennis applications.通过聚合深度网络和低秩主动学习生成人类面部动画及其在乒乓球应用中的应用
Sci Rep. 2025 Aug 1;15(1):28169. doi: 10.1038/s41598-025-13779-6.