• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

言语存在时情绪面部表情的动作单元模型

Action Unit Models of Facial Expression of Emotion in the Presence of Speech.

作者信息

Shah Miraj, Cooper David G, Cao Houwei, Gur Ruben C, Nenkova Ani, Verma Ragini

机构信息

Section of Biomedical Image Analysis, Department of Radiology, University of Pennsylvania, Philadelphia, PA19104, United States.

Department of Psychiatry, University of Pennsylvania, Philadelphia, PA19104, United States.

出版信息

Int Conf Affect Comput Intell Interact Workshops. 2013 Sep;2013:49-54. doi: 10.1109/ACII.2013.15.

DOI:10.1109/ACII.2013.15
PMID:25525561
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC4267560/
Abstract

Automatic recognition of emotion using facial expressions in the presence of speech poses a unique challenge because talking reveals clues for the affective state of the speaker but distorts the canonical expression of emotion on the face. We introduce a corpus of acted emotion expression where speech is either present (talking) or absent (silent). The corpus is uniquely suited for analysis of the interplay between the two conditions. We use a multimodal decision level fusion classifier to combine models of emotion from talking and silent faces as well as from audio to recognize five basic emotions: anger, disgust, fear, happy and sad. Our results strongly indicate that emotion prediction in the presence of speech from action unit facial features is less accurate when the person is talking. Modeling talking and silent expressions separately and fusing the two models greatly improves accuracy of prediction in the talking setting. The advantages are most pronounced when silent and talking face models are fused with predictions from audio features. In this multi-modal prediction both the combination of modalities and the separate models of talking and silent facial expression of emotion contribute to the improvement.

摘要

在存在语音的情况下利用面部表情自动识别情绪带来了独特的挑战,因为说话会揭示说话者情感状态的线索,但会扭曲脸上情绪的典型表达。我们引入了一个表演情绪表达语料库,其中语音要么存在(说话)要么不存在(沉默)。该语料库特别适合分析这两种情况之间的相互作用。我们使用多模态决策级融合分类器来结合来自说话和沉默面部以及音频的情绪模型,以识别五种基本情绪:愤怒、厌恶、恐惧、快乐和悲伤。我们的结果有力地表明,当人在说话时,基于动作单元面部特征的语音存在时的情绪预测不太准确。分别对说话和沉默表情进行建模并融合这两个模型,大大提高了说话场景下的预测准确性。当沉默和说话面部模型与音频特征的预测融合时,优势最为明显。在这种多模态预测中,模态的组合以及说话和沉默面部情绪表达的单独模型都有助于提高准确性。

相似文献

1
Action Unit Models of Facial Expression of Emotion in the Presence of Speech.言语存在时情绪面部表情的动作单元模型
Int Conf Affect Comput Intell Interact Workshops. 2013 Sep;2013:49-54. doi: 10.1109/ACII.2013.15.
2
CREMA-D: Crowd-sourced Emotional Multimodal Actors Dataset.CREMA-D:众包情感多模态演员数据集。
IEEE Trans Affect Comput. 2014 Oct-Dec;5(4):377-390. doi: 10.1109/TAFFC.2014.2336244.
3
Talking Face Generation With Audio-Deduced Emotional Landmarks.基于音频提取的情感地标进行人脸对话生成。
IEEE Trans Neural Netw Learn Syst. 2024 Oct;35(10):14099-14111. doi: 10.1109/TNNLS.2023.3274676. Epub 2024 Oct 7.
4
Evidence for shared deficits in identifying emotions from faces and from voices in autism spectrum disorders and specific language impairment.自闭症谱系障碍和特定语言障碍患者在通过面部和声音识别情绪方面存在共同缺陷的证据。
Int J Lang Commun Disord. 2015 Jul;50(4):452-66. doi: 10.1111/1460-6984.12146. Epub 2015 Jan 14.
5
Multi-modal emotion recognition using EEG and speech signals.基于脑电和语音信号的多模态情感识别。
Comput Biol Med. 2022 Oct;149:105907. doi: 10.1016/j.compbiomed.2022.105907. Epub 2022 Jul 22.
6
Mapping the emotional face. How individual face parts contribute to successful emotion recognition.绘制情感面孔。面部各个部分如何助力成功的情绪识别。
PLoS One. 2017 May 11;12(5):e0177239. doi: 10.1371/journal.pone.0177239. eCollection 2017.
7
Multimodal Emotion Detection via Attention-Based Fusion of Extracted Facial and Speech Features.基于提取的面部和语音特征的注意力融合的多模态情感检测。
Sensors (Basel). 2023 Jun 9;23(12):5475. doi: 10.3390/s23125475.
8
Common cues to emotion in the dynamic facial expressions of speech and song.言语和歌曲动态面部表情中常见的情感线索。
Q J Exp Psychol (Hove). 2015;68(5):952-70. doi: 10.1080/17470218.2014.971034. Epub 2014 Nov 25.
9
Can an anger face also be scared? Malleability of facial expressions.愤怒的表情也能表现出恐惧吗?面部表情的可延展性。
Emotion. 2012 Oct;12(5):919-25. doi: 10.1037/a0026119. Epub 2011 Nov 7.
10
Realization of Self-Adaptive Higher Teaching Management Based Upon Expression and Speech Multimodal Emotion Recognition.基于表情与语音多模态情感识别的自适应高等教学管理实现
Front Psychol. 2022 Mar 28;13:857924. doi: 10.3389/fpsyg.2022.857924. eCollection 2022.

引用本文的文献

1
Facial expressions and speech acts: experimental evidences on the role of the upper face as an illocutionary force indicating device in language comprehension.面部表情与言语行为:关于上半脸在语言理解中作为一种语力指示手段所起作用的实验证据
Cogn Process. 2017 Aug;18(3):285-306. doi: 10.1007/s10339-017-0809-6. Epub 2017 Apr 22.

本文引用的文献

1
Automated Facial Action Coding System for dynamic analysis of facial expressions in neuropsychiatric disorders.用于神经精神障碍患者面部表情动态分析的自动化面部动作编码系统。
J Neurosci Methods. 2011 Sep 15;200(2):237-56. doi: 10.1016/j.jneumeth.2011.06.023. Epub 2011 Jun 29.
2
A survey of affect recognition methods: audio, visual, and spontaneous expressions.情感识别方法综述:音频、视觉与自发表情
IEEE Trans Pattern Anal Mach Intell. 2009 Jan;31(1):39-58. doi: 10.1109/TPAMI.2008.52.
3
Validation of affective and neutral sentence content for prosodic testing.情感和中性句子内容用于韵律测试的验证。
Behav Res Methods. 2008 Nov;40(4):935-9. doi: 10.3758/BRM.40.4.935.