• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

汉英跨文化交际中多模态情感感知与生理信号融合分析:融入自注意力增强的Transformer方法

Analysis of the fusion of multimodal sentiment perception and physiological signals in Chinese-English cross-cultural communication: Transformer approach incorporating self-attention enhancement.

作者信息

Bi Xin, Zhang Tian

机构信息

School of Literature, Heilongjiang University, Harbin, Heilongjiang, China.

Department of Languages and Literary Studies, Lafayette College, Easton, PA, United States.

出版信息

PeerJ Comput Sci. 2025 May 23;11:e2890. doi: 10.7717/peerj-cs.2890. eCollection 2025.

DOI:10.7717/peerj-cs.2890
PMID:40567693
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12192752/
Abstract

With the acceleration of globalization, cross-cultural communication has become a crucial issue in various fields. Emotion, as an essential component of communication, plays a key role in improving understanding and interaction efficiency across different cultures. However, accurately recognizing emotions across cultural backgrounds remains a major challenge in affective computing, particularly due to limitations in multimodal feature fusion and temporal dependency modeling in traditional approaches. To address this, we propose the TAF-ATRM framework, which integrates Transformer and multi-head attention mechanisms for cross-cultural emotion recognition. Specifically, the framework employs bidirectional encoder representations from transformers (BERT) for semantic feature extraction from text, Mel-frequency Cepstral Coefficients (MFCC) and Residual Neural Network (ResNet) for capturing critical features from speech and facial expressions, respectively, thereby enhancing multimodal emotion recognition capability. To improve the fusion of multimodal data, the Transformer is utilized for temporal feature modeling, while multi-head attention reinforces feature representation by capturing complex inter-modal dependencies. The framework is evaluated on the MOSI and MOSEI datasets, where experimental results demonstrate that TAF-ATRM outperforms traditional methods in emotion classification accuracy and robustness, particularly in cross-cultural emotion recognition tasks. This study provides a strong technical foundation for future advancements in multimodal emotion analysis and cross-cultural affective computing.

摘要

随着全球化的加速,跨文化交流已成为各个领域的关键问题。情感作为交流的重要组成部分,在增进不同文化间的理解和互动效率方面发挥着关键作用。然而,在情感计算中,跨文化背景下准确识别情感仍然是一项重大挑战,尤其是由于传统方法在多模态特征融合和时间依赖性建模方面存在局限性。为解决这一问题,我们提出了TAF-ATRM框架,该框架集成了Transformer和多头注意力机制用于跨文化情感识别。具体而言,该框架采用来自Transformer的双向编码器表示(BERT)从文本中提取语义特征,分别使用梅尔频率倒谱系数(MFCC)和残差神经网络(ResNet)从语音和面部表情中捕捉关键特征,从而增强多模态情感识别能力。为了改进多模态数据的融合,Transformer用于时间特征建模,而多头注意力通过捕捉复杂的模态间依赖性来强化特征表示。该框架在MOSI和MOSEI数据集上进行了评估,实验结果表明TAF-ATRM在情感分类准确性和鲁棒性方面优于传统方法,尤其是在跨文化情感识别任务中。本研究为多模态情感分析和跨文化情感计算的未来发展提供了坚实的技术基础。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aa43/12192752/4f267786d3f8/peerj-cs-11-2890-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aa43/12192752/2ac9e9b4778a/peerj-cs-11-2890-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aa43/12192752/eb7b30b3ea05/peerj-cs-11-2890-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aa43/12192752/8f87b4e908fc/peerj-cs-11-2890-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aa43/12192752/3a540c535c12/peerj-cs-11-2890-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aa43/12192752/73ee710e3d86/peerj-cs-11-2890-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aa43/12192752/182447eec41d/peerj-cs-11-2890-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aa43/12192752/9c8b58a00408/peerj-cs-11-2890-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aa43/12192752/3f37479a6556/peerj-cs-11-2890-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aa43/12192752/4f267786d3f8/peerj-cs-11-2890-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aa43/12192752/2ac9e9b4778a/peerj-cs-11-2890-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aa43/12192752/eb7b30b3ea05/peerj-cs-11-2890-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aa43/12192752/8f87b4e908fc/peerj-cs-11-2890-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aa43/12192752/3a540c535c12/peerj-cs-11-2890-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aa43/12192752/73ee710e3d86/peerj-cs-11-2890-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aa43/12192752/182447eec41d/peerj-cs-11-2890-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aa43/12192752/9c8b58a00408/peerj-cs-11-2890-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aa43/12192752/3f37479a6556/peerj-cs-11-2890-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aa43/12192752/4f267786d3f8/peerj-cs-11-2890-g009.jpg

相似文献

1
Analysis of the fusion of multimodal sentiment perception and physiological signals in Chinese-English cross-cultural communication: Transformer approach incorporating self-attention enhancement.汉英跨文化交际中多模态情感感知与生理信号融合分析:融入自注意力增强的Transformer方法
PeerJ Comput Sci. 2025 May 23;11:e2890. doi: 10.7717/peerj-cs.2890. eCollection 2025.
2
Trajectory-Ordered Objectives for Self-Supervised Representation Learning of Temporal Healthcare Data Using Transformers: Model Development and Evaluation Study.使用Transformer进行时间序列医疗数据自监督表示学习的轨迹有序目标:模型开发与评估研究
JMIR Med Inform. 2025 Jun 4;13:e68138. doi: 10.2196/68138.
3
Cross-corpus speech emotion recognition with transformers: Leveraging handcrafted features and data augmentation.基于 Transformer 的跨语料库语音情感识别:利用手工特征和数据增强。
Comput Biol Med. 2024 Sep;179:108841. doi: 10.1016/j.compbiomed.2024.108841. Epub 2024 Jul 12.
4
Exploring the Potential of Electroencephalography Signal-Based Image Generation Using Diffusion Models: Integrative Framework Combining Mixed Methods and Multimodal Analysis.利用扩散模型探索基于脑电图信号的图像生成潜力:结合混合方法和多模态分析的综合框架
JMIR Med Inform. 2025 Jun 25;13:e72027. doi: 10.2196/72027.
5
UMEDNet: a multimodal approach for emotion detection in the Urdu language.UMEDNet:一种用于乌尔都语情感检测的多模态方法。
PeerJ Comput Sci. 2025 May 1;11:e2861. doi: 10.7717/peerj-cs.2861. eCollection 2025.
6
DGCFNet: Dual Global Context Fusion Network for remote sensing image semantic segmentation.DGCFNet:用于遥感图像语义分割的双全局上下文融合网络
PeerJ Comput Sci. 2025 Mar 27;11:e2786. doi: 10.7717/peerj-cs.2786. eCollection 2025.
7
Detecting Redundant Health Survey Questions by Using Language-Agnostic Bidirectional Encoder Representations From Transformers Sentence Embedding: Algorithm Development Study.使用来自Transformer句子嵌入的语言无关双向编码器表示法检测冗余健康调查问题:算法开发研究
JMIR Med Inform. 2025 Jun 10;13:e71687. doi: 10.2196/71687.
8
[Research on bimodal emotion recognition algorithm based on multi-branch bidirectional multi-scale time perception].基于多分支双向多尺度时间感知的双峰情感识别算法研究
Sheng Wu Yi Xue Gong Cheng Xue Za Zhi. 2025 Jun 25;42(3):528-536. doi: 10.7507/1001-5515.202404047.
9
DashFusion: Dual-Stream Alignment With Hierarchical Bottleneck Fusion for Multimodal Sentiment Analysis.DashFusion:用于多模态情感分析的具有分层瓶颈融合的双流对齐
IEEE Trans Neural Netw Learn Syst. 2025 Jun 18;PP. doi: 10.1109/TNNLS.2025.3578618.
10
TLTNet: A novel transscale cascade layered transformer network for enhanced retinal blood vessel segmentation.TLTNet:一种新颖的跨尺度级联分层Transformer 网络,用于增强视网膜血管分割。
Comput Biol Med. 2024 Aug;178:108773. doi: 10.1016/j.compbiomed.2024.108773. Epub 2024 Jun 25.

本文引用的文献

1
Adaptive Spatial-Temporal Aware Graph Learning for EEG-Based Emotion Recognition.用于基于脑电图的情绪识别的自适应时空感知图学习
Cyborg Bionic Syst. 2024 May 8;5:0088. doi: 10.34133/cbsystems.0088. eCollection 2024.
2
EEG-based emotion recognition with autoencoder feature fusion and MSC-TimesNet model.基于脑电图的情感识别与自动编码器特征融合及MSC-TimesNet模型
Comput Methods Biomech Biomed Engin. 2025 Mar 17:1-18. doi: 10.1080/10255842.2025.2477801.
3
Toward the enhancement of affective brain-computer interfaces using dependence within EEG series.
利用脑电图系列中的相关性增强情感脑机接口。
J Neural Eng. 2025 Apr 1;22(2). doi: 10.1088/1741-2552/adbfc0.
4
TalkingStyle: Personalized Speech-Driven 3D Facial Animation with Style Preservation.谈话风格:具有风格保留的个性化语音驱动3D面部动画
IEEE Trans Vis Comput Graph. 2024 Jun 11;PP. doi: 10.1109/TVCG.2024.3409568.
5
ResNet and its application to medical image processing: Research progress and challenges.ResNet 及其在医学图像处理中的应用:研究进展与挑战。
Comput Methods Programs Biomed. 2023 Oct;240:107660. doi: 10.1016/j.cmpb.2023.107660. Epub 2023 Jun 8.
6
Robust Multimodal Emotion Recognition from Conversation with Transformer-Based Crossmodality Fusion.基于 Transformer 的跨模态融合的对话中的稳健多模态情感识别。
Sensors (Basel). 2021 Jul 19;21(14):4913. doi: 10.3390/s21144913.
7
Multi-attention Recurrent Network for Human Communication Comprehension.用于人类交流理解的多注意力循环网络。
Proc AAAI Conf Artif Intell. 2018 Feb;2018:5642-5649.
8
Words Can Shift: Dynamically Adjusting Word Representations Using Nonverbal Behaviors.词语可变换:利用非语言行为动态调整词语表征
Proc AAAI Conf Artif Intell. 2019 Jul;33(1):7216-7223.