• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

AFR-BERT:基于注意力机制的特征相关融合多模态情感分析模型。

AFR-BERT: Attention-based mechanism feature relevance fusion multimodal sentiment analysis model.

机构信息

Department of Software Engineering, Faculty of Information and Computer Engineering, The Northeast Forestry University, Harbin, China.

出版信息

PLoS One. 2022 Sep 9;17(9):e0273936. doi: 10.1371/journal.pone.0273936. eCollection 2022.

DOI:10.1371/journal.pone.0273936
PMID:36084041
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9462790/
Abstract

Multimodal sentiment analysis is an essential task in natural language processing which refers to the fact that machines can analyze and recognize emotions through logical reasoning and mathematical operations after learning multimodal emotional features. For the problem of how to consider the effective fusion of multimodal data and the relevance of multimodal data in multimodal sentiment analysis, we propose an attention-based mechanism feature relevance fusion multimodal sentiment analysis model (AFR-BERT). In the data pre-processing stage, text features are extracted using the pre-trained language model BERT (Bi-directional Encoder Representation from Transformers), and the BiLSTM (Bi-directional Long Short-Term Memory) is used to obtain the internal information of the audio. In the data fusion phase, the multimodal data fusion network effectively fuses multimodal features through the interaction of text and audio information. During the data analysis phase, the multimodal data association network analyzes the data by exploring the correlation of fused information between text and audio. In the data output phase, the model outputs the results of multimodal sentiment analysis. We conducted extensive comparative experiments on the publicly available sentiment analysis datasets CMU-MOSI and CMU-MOSEI. The experimental results show that AFR-BERT improves on the classical multimodal sentiment analysis model in terms of relevant performance metrics. In addition, ablation experiments and example analysis show that the multimodal data analysis network in AFR-BERT can effectively capture and analyze the sentiment features in text and audio.

摘要

多模态情感分析是自然语言处理中的一项重要任务,指的是机器经过学习多模态情感特征后,能够通过逻辑推理和数学运算分析和识别情感。针对多模态情感分析中如何考虑多模态数据的有效融合以及多模态数据的相关性问题,提出了一种基于注意力机制特征相关性融合的多模态情感分析模型(AFR-BERT)。在数据预处理阶段,使用预训练语言模型 BERT(来自 Transformer 的双向编码器表示)提取文本特征,并使用 BiLSTM(双向长短期记忆)获取音频的内部信息。在数据融合阶段,多模态数据融合网络通过文本和音频信息的交互,有效地融合多模态特征。在数据分析阶段,多模态数据关联网络通过探索文本和音频之间融合信息的相关性来分析数据。在数据输出阶段,模型输出多模态情感分析的结果。在公开的情感分析数据集 CMU-MOSI 和 CMU-MOSEI 上进行了广泛的对比实验。实验结果表明,AFR-BERT 在相关性能指标上优于经典的多模态情感分析模型。此外,消融实验和案例分析表明,AFR-BERT 中的多模态数据分析网络能够有效地捕捉和分析文本和音频中的情感特征。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6d72/9462790/264aff26a6b4/pone.0273936.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6d72/9462790/c0b87a88a249/pone.0273936.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6d72/9462790/eab3464d807d/pone.0273936.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6d72/9462790/264aff26a6b4/pone.0273936.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6d72/9462790/c0b87a88a249/pone.0273936.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6d72/9462790/eab3464d807d/pone.0273936.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6d72/9462790/264aff26a6b4/pone.0273936.g003.jpg

相似文献

1
AFR-BERT: Attention-based mechanism feature relevance fusion multimodal sentiment analysis model.AFR-BERT:基于注意力机制的特征相关融合多模态情感分析模型。
PLoS One. 2022 Sep 9;17(9):e0273936. doi: 10.1371/journal.pone.0273936. eCollection 2022.
2
A BERT Framework to Sentiment Analysis of Tweets.一种用于推文情感分析的BERT框架。
Sensors (Basel). 2023 Jan 2;23(1):506. doi: 10.3390/s23010506.
3
Multimodal Sentiment Analysis Based on Cross-Modal Attention and Gated Cyclic Hierarchical Fusion Networks.基于跨模态注意力和门控循环层次融合网络的多模态情感分析。
Comput Intell Neurosci. 2022 Aug 9;2022:4767437. doi: 10.1155/2022/4767437. eCollection 2022.
4
Integrating Multimodal Information in Large Pretrained Transformers.在大型预训练变压器中整合多模态信息。
Proc Conf Assoc Comput Linguist Meet. 2020 Jul;2020:2359-2369. doi: 10.18653/v1/2020.acl-main.214.
5
Transfer Learning for Sentiment Analysis Using BERT Based Supervised Fine-Tuning.基于 BERT 的有监督微调的情感分析中的迁移学习。
Sensors (Basel). 2022 May 30;22(11):4157. doi: 10.3390/s22114157.
6
Research on sentiment classification for netizens based on the BERT-BiLSTM-TextCNN model.基于BERT-BiLSTM-TextCNN模型的网民情感分类研究
PeerJ Comput Sci. 2022 Jun 8;8:e1005. doi: 10.7717/peerj-cs.1005. eCollection 2022.
7
AB-GRU: An attention-based bidirectional GRU model for multimodal sentiment fusion and analysis.AB-GRU:一种用于多模态情感融合与分析的基于注意力机制的双向门控循环单元模型。
Math Biosci Eng. 2023 Sep 27;20(10):18523-18544. doi: 10.3934/mbe.2023822.
8
Dynamic Invariant-Specific Representation Fusion Network for Multimodal Sentiment Analysis.动态不变特定表示融合网络用于多模态情感分析。
Comput Intell Neurosci. 2022 Jan 24;2022:2105593. doi: 10.1155/2022/2105593. eCollection 2022.
9
Text Sentiment Classification Based on BERT Embedding and Sliced Multi-Head Self-Attention Bi-GRU.基于 BERT 嵌入和切片多头自注意力 Bi-GRU 的文本情感分类
Sensors (Basel). 2023 Jan 28;23(3):1481. doi: 10.3390/s23031481.
10
Multi-class sentiment analysis of urdu text using multilingual BERT.使用多语言 BERT 进行乌尔都语文本的多类情感分析。
Sci Rep. 2022 Mar 31;12(1):5436. doi: 10.1038/s41598-022-09381-9.

引用本文的文献

1
Efficient screening of pharmacological broad-spectrum anti-cancer peptides utilizing advanced bidirectional Encoder representation from Transformers strategy.利用先进的基于变换器的双向编码器表征策略高效筛选药理广谱抗癌肽。
Heliyon. 2024 May 1;10(9):e30373. doi: 10.1016/j.heliyon.2024.e30373. eCollection 2024 May 15.
2
BERT-based language model for accurate drug adverse event extraction from social media: implementation, evaluation, and contributions to pharmacovigilance practices.基于 BERT 的社交媒体中药物不良反应精准提取语言模型:实现、评估及对药物警戒实践的贡献。
Front Public Health. 2024 Apr 23;12:1392180. doi: 10.3389/fpubh.2024.1392180. eCollection 2024.
3

本文引用的文献

1
Integrating Multimodal Information in Large Pretrained Transformers.在大型预训练变压器中整合多模态信息。
Proc Conf Assoc Comput Linguist Meet. 2020 Jul;2020:2359-2369. doi: 10.18653/v1/2020.acl-main.214.
2
Multimodal Transformer for Unaligned Multimodal Language Sequences.用于未对齐多模态语言序列的多模态变换器
Proc Conf Assoc Comput Linguist Meet. 2019 Jul;2019:6558-6569. doi: 10.18653/v1/p19-1656.
3
Multi-attention Recurrent Network for Human Communication Comprehension.用于人类交流理解的多注意力循环网络。
Application of the transformer model algorithm in chinese word sense disambiguation: a case study in chinese language.
变压器模型算法在中文词义消歧中的应用:以中文语言为例的研究
Sci Rep. 2024 Mar 15;14(1):6320. doi: 10.1038/s41598-024-56976-5.
4
Automatic literature screening using the PAJO deep-learning model for clinical practice guidelines.使用 PAJO 深度学习模型进行自动文献筛选以制定临床实践指南。
BMC Med Inform Decis Mak. 2023 Nov 3;23(1):247. doi: 10.1186/s12911-023-02328-8.
Proc AAAI Conf Artif Intell. 2018 Feb;2018:5642-5649.
4
Words Can Shift: Dynamically Adjusting Word Representations Using Nonverbal Behaviors.词语可变换:利用非语言行为动态调整词语表征
Proc AAAI Conf Artif Intell. 2019 Jul;33(1):7216-7223.
5
Distinguishing anger and anxiety in terms of emotional response factors.从情绪反应因素方面区分愤怒和焦虑。
J Consult Clin Psychol. 1974 Feb;42(1):79-83. doi: 10.1037/h0035915.