• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

社交网络舆情中不平衡样本数据的多模态特征融合方法

Multimodal Feature Fusion Method for Unbalanced Sample Data in Social Network Public Opinion.

机构信息

School of Cyber Security, Changchun University, Changchun 130022, China.

School of Computer Science and Technology, Changchun University, Changchun 130022, China.

出版信息

Sensors (Basel). 2022 Jul 25;22(15):5528. doi: 10.3390/s22155528.

DOI:10.3390/s22155528
PMID:35898032
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9331324/
Abstract

With the wide application of social media, public opinion analysis in social networks has been unable to be met through text alone because the existing public opinion information includes data information of various modalities, such as voice, text, and facial expressions. Therefore multi-modal emotion analysis is the current focus of public opinion analysis. In addition, multi-modal emotion recognition of speech is an important factor restricting the multi-modal emotion analysis. In this paper, the emotion feature retrieval method for speech is firstly explored and the processing method of sample disequilibrium data is then analyzed. By comparing and studying the different feature fusion methods of text and speech, respectively, the multi-modal feature fusion method for sample disequilibrium data is proposed to realize multi-modal emotion recognition. Experiments are performed using two publicly available datasets (IEMOCAP and MELD), which shows that processing multi-modality data through this method can obtain good fine-grained emotion recognition results, laying a foundation for subsequent social public opinion analysis.

摘要

随着社交媒体的广泛应用,仅通过文本进行网络社交舆情分析已经无法满足需求,因为现有的舆情信息包含了语音、文本和表情等多种模态的数据信息。因此,多模态情感分析是当前舆情分析的重点。此外,语音的多模态情感识别是制约多模态情感分析的一个重要因素。本文首先探索了语音的情感特征检索方法,然后分析了样本不平衡数据的处理方法。通过分别比较和研究文本和语音的不同特征融合方法,提出了一种用于样本不平衡数据的多模态特征融合方法,以实现多模态情感识别。实验分别使用了两个公开可用的数据集(IEMOCAP 和 MELD),结果表明,通过这种方法处理多模态数据可以获得良好的细粒度情感识别结果,为后续的社会舆情分析奠定了基础。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/1c4f8cdd0488/sensors-22-05528-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/3391f823a6de/sensors-22-05528-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/68e04b078d61/sensors-22-05528-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/c309afa61676/sensors-22-05528-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/27434e624fad/sensors-22-05528-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/e86b4d3358b4/sensors-22-05528-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/20f3c1eac3b8/sensors-22-05528-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/0b896554d90b/sensors-22-05528-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/73a56257bb98/sensors-22-05528-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/95e4055571d2/sensors-22-05528-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/2f117e255404/sensors-22-05528-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/66304b49357c/sensors-22-05528-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/f0a66c2197c8/sensors-22-05528-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/1c4f8cdd0488/sensors-22-05528-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/3391f823a6de/sensors-22-05528-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/68e04b078d61/sensors-22-05528-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/c309afa61676/sensors-22-05528-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/27434e624fad/sensors-22-05528-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/e86b4d3358b4/sensors-22-05528-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/20f3c1eac3b8/sensors-22-05528-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/0b896554d90b/sensors-22-05528-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/73a56257bb98/sensors-22-05528-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/95e4055571d2/sensors-22-05528-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/2f117e255404/sensors-22-05528-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/66304b49357c/sensors-22-05528-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/f0a66c2197c8/sensors-22-05528-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/da08/9331324/1c4f8cdd0488/sensors-22-05528-g013.jpg

相似文献

1
Multimodal Feature Fusion Method for Unbalanced Sample Data in Social Network Public Opinion.社交网络舆情中不平衡样本数据的多模态特征融合方法
Sensors (Basel). 2022 Jul 25;22(15):5528. doi: 10.3390/s22155528.
2
Multi-Modal Fusion Emotion Recognition Method of Speech Expression Based on Deep Learning.基于深度学习的语音表达多模态融合情感识别方法
Front Neurorobot. 2021 Jul 9;15:697634. doi: 10.3389/fnbot.2021.697634. eCollection 2021.
3
Multimodal transformer augmented fusion for speech emotion recognition.用于语音情感识别的多模态变压器增强融合
Front Neurorobot. 2023 May 22;17:1181598. doi: 10.3389/fnbot.2023.1181598. eCollection 2023.
4
GCF-Net: global-aware cross-modal feature fusion network for speech emotion recognition.GCF-Net:用于语音情感识别的全局感知跨模态特征融合网络
Front Neurosci. 2023 May 4;17:1183132. doi: 10.3389/fnins.2023.1183132. eCollection 2023.
5
Multimodal Emotion Detection via Attention-Based Fusion of Extracted Facial and Speech Features.基于提取的面部和语音特征的注意力融合的多模态情感检测。
Sensors (Basel). 2023 Jun 9;23(12):5475. doi: 10.3390/s23125475.
6
Multi-modal emotion recognition using EEG and speech signals.基于脑电和语音信号的多模态情感识别。
Comput Biol Med. 2022 Oct;149:105907. doi: 10.1016/j.compbiomed.2022.105907. Epub 2022 Jul 22.
7
Multimodal Emotion Recognition Based on Cascaded Multichannel and Hierarchical Fusion.基于级联多通道和分层融合的多模态情绪识别。
Comput Intell Neurosci. 2023 Jan 5;2023:9645611. doi: 10.1155/2023/9645611. eCollection 2023.
8
Emotional Speech Recognition Using Deep Neural Networks.使用深度神经网络进行情感语音识别。
Sensors (Basel). 2022 Feb 12;22(4):1414. doi: 10.3390/s22041414.
9
Research on cross-modal emotion recognition based on multi-layer semantic fusion.基于多层语义融合的跨模态情感识别研究
Math Biosci Eng. 2024 Jan 17;21(2):2488-2514. doi: 10.3934/mbe.2024110.
10
Multi-Modal Residual Perceptron Network for Audio-Video Emotion Recognition.多模态残差感知机网络的音视频情感识别。
Sensors (Basel). 2021 Aug 12;21(16):5452. doi: 10.3390/s21165452.

引用本文的文献

1
AVaTER: Fusing Audio, Visual, and Textual Modalities Using Cross-Modal Attention for Emotion Recognition.AVaTER:使用跨模态注意力融合音频、视觉和文本模态进行情感识别。
Sensors (Basel). 2024 Sep 10;24(18):5862. doi: 10.3390/s24185862.
2
Enhancing Human Activity Recognition through Integrated Multimodal Analysis: A Focus on RGB Imaging, Skeletal Tracking, and Pose Estimation.通过集成多模态分析增强人类活动识别:重点关注 RGB 成像、骨骼跟踪和姿势估计。
Sensors (Basel). 2024 Jul 17;24(14):4646. doi: 10.3390/s24144646.
3
Fake News Detection Model on Social Media by Leveraging Sentiment Analysis of News Content and Emotion Analysis of Users' Comments.

本文引用的文献

1
Acoustic Sensor-Based Soundscape Analysis and Acoustic Assessment of Bird Species Richness in Shennongjia National Park, China.基于声学传感器的中国神农架国家公园鸟鸣声景观分析与鸟类物种多样性声学评估。
Sensors (Basel). 2022 May 28;22(11):4117. doi: 10.3390/s22114117.
2
Topic-Emotion Propagation Mechanism of Public Emergencies in Social Networks.主题-社交网络中公共突发事件的情感传播机制。
Sensors (Basel). 2021 Jul 1;21(13):4516. doi: 10.3390/s21134516.
3
Conversational Memory Network for Emotion Recognition in Dyadic Dialogue Videos.用于二元对话视频中情感识别的对话记忆网络
社交媒体上的假新闻检测模型,利用新闻内容的情感分析和用户评论的情绪分析。
Sensors (Basel). 2023 Feb 4;23(4):1748. doi: 10.3390/s23041748.
Proc Conf. 2018 Jun;2018:2122-2132. doi: 10.18653/v1/n18-1193.
4
Multimodal emotion recognition using EEG and eye tracking data.利用脑电图(EEG)和眼动追踪数据进行多模态情感识别。
Annu Int Conf IEEE Eng Med Biol Soc. 2014;2014:5040-3. doi: 10.1109/EMBC.2014.6944757.