• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

MCCA-VNet:一种基于视觉变换器(Vit)的、基于面部编码的微表情识别深度学习方法。

MCCA-VNet: A Vit-Based Deep Learning Approach for Micro-Expression Recognition Based on Facial Coding.

作者信息

Zhang Dehao, Zhang Tao, Sun Haijiang, Tang Yanhui, Liu Qiaoyuan

机构信息

Changchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun 130033, China.

University of Chinese Academy of Sciences, Beijing 100049, China.

出版信息

Sensors (Basel). 2024 Nov 26;24(23):7549. doi: 10.3390/s24237549.

DOI:10.3390/s24237549
PMID:39686086
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11644463/
Abstract

In terms of facial expressions, micro-expressions are more realistic than macro-expressions and provide more valuable information, which can be widely used in psychological counseling and clinical diagnosis. In the past few years, deep learning methods based on optical flow and Transformer have achieved excellent results in this field, but most of the current algorithms are mainly concentrated on establishing a serialized token through the self-attention model, and they do not take into account the spatial relationship between facial landmarks. For the locality and changes in the micro-facial conditions themselves, we propose the deep learning model MCCA-VNET on the basis of Transformer. We effectively extract the changing features as the input of the model, fusing channel attention and spatial attention into Vision Transformer to capture correlations between features in different dimensions, which enhances the accuracy of the identification of micro-expressions. In order to verify the effectiveness of the algorithm mentioned, we conduct experimental testing in the SAMM, CAS (ME) II, and SMIC datasets and compared the results with other former best algorithms. Our algorithms can improve the UF1 score and UAR score to, respectively, 0.8676 and 0.8622 for the composite dataset, and they are better than other algorithms on multiple indicators, achieving the best comprehensive performance.

摘要

在面部表情方面,微表情比宏表情更真实,能提供更有价值的信息,可广泛应用于心理咨询和临床诊断。在过去几年中,基于光流和Transformer的深度学习方法在该领域取得了优异成果,但目前大多数算法主要集中于通过自注意力模型建立序列化令牌,并未考虑面部关键点之间的空间关系。针对微面部状况本身的局部性和变化,我们在Transformer的基础上提出了深度学习模型MCCA-VNET。我们有效地提取变化特征作为模型输入,将通道注意力和空间注意力融合到视觉Transformer中,以捕捉不同维度特征之间的相关性,从而提高微表情识别的准确率。为验证上述算法的有效性,我们在SAMM、CAS(ME)II和SMIC数据集中进行了实验测试,并将结果与其他先前的最佳算法进行比较。对于综合数据集,我们的算法可将UF1分数和UAR分数分别提高到0.8676和0.8622,在多个指标上优于其他算法,实现了最佳综合性能。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d259/11644463/c9ab10cabed2/sensors-24-07549-g007a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d259/11644463/e55edd8f82b6/sensors-24-07549-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d259/11644463/b8637b5524f5/sensors-24-07549-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d259/11644463/bd4557b33c71/sensors-24-07549-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d259/11644463/26318b6c3f0d/sensors-24-07549-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d259/11644463/ca5054e8cbd0/sensors-24-07549-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d259/11644463/a1965b67ba96/sensors-24-07549-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d259/11644463/c9ab10cabed2/sensors-24-07549-g007a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d259/11644463/e55edd8f82b6/sensors-24-07549-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d259/11644463/b8637b5524f5/sensors-24-07549-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d259/11644463/bd4557b33c71/sensors-24-07549-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d259/11644463/26318b6c3f0d/sensors-24-07549-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d259/11644463/ca5054e8cbd0/sensors-24-07549-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d259/11644463/a1965b67ba96/sensors-24-07549-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d259/11644463/c9ab10cabed2/sensors-24-07549-g007a.jpg

相似文献

1
MCCA-VNet: A Vit-Based Deep Learning Approach for Micro-Expression Recognition Based on Facial Coding.MCCA-VNet:一种基于视觉变换器(Vit)的、基于面部编码的微表情识别深度学习方法。
Sensors (Basel). 2024 Nov 26;24(23):7549. doi: 10.3390/s24237549.
2
Two-Level Spatio-Temporal Feature Fused Two-Stream Network for Micro-Expression Recognition.基于两级时空特征融合双流网络的微表情识别方法。
Sensors (Basel). 2024 Feb 29;24(5):1574. doi: 10.3390/s24051574.
3
Facial Micro-Expression Recognition Enhanced by Score Fusion and a Hybrid Model from Convolutional LSTM and Vision Transformer.基于卷积长短期记忆网络和视觉Transformer融合的评分融合与混合模型增强的面部微表情识别。
Sensors (Basel). 2023 Jun 16;23(12):5650. doi: 10.3390/s23125650.
4
CDGT: Constructing diverse graph transformers for emotion recognition from facial videos.构建用于面部视频情感识别的多样化图变换模型。
Neural Netw. 2024 Nov;179:106573. doi: 10.1016/j.neunet.2024.106573. Epub 2024 Jul 25.
5
A fine-grained human facial key feature extraction and fusion method for emotion recognition.一种用于情感识别的细粒度人类面部关键特征提取与融合方法。
Sci Rep. 2025 Feb 20;15(1):6153. doi: 10.1038/s41598-025-90440-2.
6
Lightweight ViT Model for Micro-Expression Recognition Enhanced by Transfer Learning.基于迁移学习增强的用于微表情识别的轻量级视觉Transformer模型
Front Neurorobot. 2022 Jun 30;16:922761. doi: 10.3389/fnbot.2022.922761. eCollection 2022.
7
Facial micro-expression recognition using stochastic graph convolutional network and dual transferred learning.基于随机图卷积网络和双重迁移学习的面部微表情识别。
Neural Netw. 2024 Oct;178:106421. doi: 10.1016/j.neunet.2024.106421. Epub 2024 Jun 3.
8
Facial Expression Recognition Based on Fine-Tuned Channel-Spatial Attention Transformer.基于精调通道-空间注意力转换器的面部表情识别。
Sensors (Basel). 2023 Jul 30;23(15):6799. doi: 10.3390/s23156799.
9
A Student Facial Expression Recognition Model Based on Multi-Scale and Deep Fine-Grained Feature Attention Enhancement.基于多尺度和深度细粒度特征注意力增强的学生面部表情识别模型。
Sensors (Basel). 2024 Oct 20;24(20):6748. doi: 10.3390/s24206748.
10
TriCAFFNet: A Tri-Cross-Attention Transformer with a Multi-Feature Fusion Network for Facial Expression Recognition.TriCAFFNet:一种具有多特征融合网络的三交叉注意力转换器,用于面部表情识别。
Sensors (Basel). 2024 Aug 21;24(16):5391. doi: 10.3390/s24165391.

本文引用的文献

1
ME-PLAN: A deep prototypical learning with local attention network for dynamic micro-expression recognition.ME-PLAN:一种基于深度原型学习和局部注意力网络的动态微表情识别方法。
Neural Netw. 2022 Sep;153:427-443. doi: 10.1016/j.neunet.2022.06.024. Epub 2022 Jun 24.
2
Video-Based Facial Micro-Expression Analysis: A Survey of Datasets, Features and Algorithms.基于视频的面部微表情分析:数据集、特征和算法综述。
IEEE Trans Pattern Anal Mach Intell. 2022 Sep;44(9):5826-5846. doi: 10.1109/TPAMI.2021.3067464. Epub 2022 Aug 4.
3
Revealing the Invisible with Model and Data Shrinking for Composite-database Micro-expression Recognition.
通过模型和数据缩减实现复合数据库微表情识别中的不可见信息揭示
IEEE Trans Image Process. 2020 Aug 26;PP. doi: 10.1109/TIP.2020.3018222.
4
Going Deeper in Spiking Neural Networks: VGG and Residual Architectures.深入探索脉冲神经网络:VGG和残差架构。
Front Neurosci. 2019 Mar 7;13:95. doi: 10.3389/fnins.2019.00095. eCollection 2019.
5
CASME II: an improved spontaneous micro-expression database and the baseline evaluation.CASME II:一个改进的自发微表情数据库及基准评估。
PLoS One. 2014 Jan 27;9(1):e86041. doi: 10.1371/journal.pone.0086041. eCollection 2014.
6
Fully automatic recognition of the temporal phases of facial actions.面部动作时间阶段的全自动识别。
IEEE Trans Syst Man Cybern B Cybern. 2012 Feb;42(1):28-43. doi: 10.1109/TSMCB.2011.2163710. Epub 2011 Sep 15.
7
Dynamic texture recognition using local binary patterns with an application to facial expressions.基于局部二值模式的动态纹理识别及其在面部表情中的应用
IEEE Trans Pattern Anal Mach Intell. 2007 Jun;29(6):915-28. doi: 10.1109/TPAMI.2007.1110.