• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

一种使用深度学习的社交媒体多模态仇恨言论检测综合框架。

A comprehensive framework for multi-modal hate speech detection in social media using deep learning.

作者信息

Prabhu R, Seethalakshmi V

机构信息

Department of Information Technology, Dr. Mahalingam College of Engineering and Technology, Pollachi, India.

Department of Electronics and Communication Engineering, KPR Institute of Engineering and Technology, Coimbatore, India.

出版信息

Sci Rep. 2025 Apr 15;15(1):13020. doi: 10.1038/s41598-025-94069-z.

DOI:10.1038/s41598-025-94069-z
PMID:40234479
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12000576/
Abstract

As social media platforms evolve, hate speech increasingly manifests across multiple modalities, including text, images, audio, and video, challenging traditional detection systems focused on single modalities. Hence, this research proposes a novel Multi-modal Hate Speech Detection Framework (MHSDF) that combines Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs) to analyze complex, heterogeneous data streams. This hybrid approach leverages CNNs for spatial feature extraction, such as identifying visual cues in images and local text patterns, and Long Short Term Memory (LSTM) for modeling temporal dependencies and sequential information in text and audio. For textual content, utilize state-of-the-art word embeddings, including Word2Vec and BERT, to capture semantic relationships and contextual nuances. The framework integrates CNNs to extract n-gram patterns and RNNs to model long-range dependency up to sequences of up to 100 tokens. CNNs extract key spatial features in visual tasks, while LSTMs process video sequences to capture evolving visual patterns. Image spatial features refer to object localization, color distributions, and text extracted via Optical Character Recognition (OCR). The fusion mechanism employs attention mechanisms to prioritize key interactions between modalities, enabling the detection of nuanced hate speech, such as memes that blend offensive imagery with implicit text, sarcastic videos where toxicity is conveyed through tone and facial expressions, and multi-layered content that embeds discriminatory meaning, across different formats. The numerical findings show that the proposed MHSDF model increases the detection accuracy ratio of 98.53%, robustness ratio of 97.64%, interpretability ratio of 97.71%, scalability ratio of 98.67%, and performance ratio of 99.21% compared to other existing models. Furthermore, the model's interpretability is enhanced through attention-based explanations, which provide insights into how multi-modal hate speech is identified. The framework improves traceability of decisions, interpretability by modality, and overall transparency.

摘要

随着社交媒体平台的发展,仇恨言论越来越多地以多种形式表现出来,包括文本、图像、音频和视频,这对专注于单一形式的传统检测系统构成了挑战。因此,本研究提出了一种新颖的多模态仇恨言论检测框架(MHSDF),该框架结合了卷积神经网络(CNN)和循环神经网络(RNN)来分析复杂的异构数据流。这种混合方法利用CNN进行空间特征提取,例如识别图像中的视觉线索和局部文本模式,并利用长短期记忆(LSTM)对文本和音频中的时间依赖性和序列信息进行建模。对于文本内容,利用包括Word2Vec和BERT在内的最新词嵌入来捕捉语义关系和上下文细微差别。该框架集成了CNN以提取n元语法模式,并集成了RNN以对长达100个词元的序列进行长距离依赖性建模。CNN在视觉任务中提取关键空间特征,而LSTM处理视频序列以捕捉不断演变的视觉模式。图像空间特征是指目标定位、颜色分布以及通过光学字符识别(OCR)提取的文本。融合机制采用注意力机制来优先处理不同模态之间的关键交互,从而能够检测细微的仇恨言论,例如将冒犯性图像与隐含文本混合的表情包、通过语气和面部表情传达毒性的讽刺视频,以及嵌入歧视性含义的多层内容,这些内容跨越不同格式。数值结果表明,与其他现有模型相比,所提出的MHSDF模型的检测准确率提高了98.53%,鲁棒性提高了97.64%,可解释性提高了97.71%,可扩展性提高了98.67%,性能提高了99.21%。此外,通过基于注意力的解释增强了模型的可解释性,这些解释提供了关于如何识别多模态仇恨言论的见解。该框架提高了决策的可追溯性、按模态的可解释性以及整体透明度。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/94a6/12000576/6328e340bd76/41598_2025_94069_Fig11_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/94a6/12000576/fa88aea22379/41598_2025_94069_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/94a6/12000576/0b1e39e59c18/41598_2025_94069_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/94a6/12000576/ac186f787eea/41598_2025_94069_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/94a6/12000576/fc3816e30941/41598_2025_94069_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/94a6/12000576/d9b9cb411300/41598_2025_94069_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/94a6/12000576/79d1479e370a/41598_2025_94069_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/94a6/12000576/6eadcd094879/41598_2025_94069_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/94a6/12000576/7f8ac9a52a5e/41598_2025_94069_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/94a6/12000576/829b93e0b40f/41598_2025_94069_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/94a6/12000576/3f6fa5ecfd71/41598_2025_94069_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/94a6/12000576/6328e340bd76/41598_2025_94069_Fig11_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/94a6/12000576/fa88aea22379/41598_2025_94069_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/94a6/12000576/0b1e39e59c18/41598_2025_94069_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/94a6/12000576/ac186f787eea/41598_2025_94069_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/94a6/12000576/fc3816e30941/41598_2025_94069_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/94a6/12000576/d9b9cb411300/41598_2025_94069_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/94a6/12000576/79d1479e370a/41598_2025_94069_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/94a6/12000576/6eadcd094879/41598_2025_94069_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/94a6/12000576/7f8ac9a52a5e/41598_2025_94069_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/94a6/12000576/829b93e0b40f/41598_2025_94069_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/94a6/12000576/3f6fa5ecfd71/41598_2025_94069_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/94a6/12000576/6328e340bd76/41598_2025_94069_Fig11_HTML.jpg

相似文献

1
A comprehensive framework for multi-modal hate speech detection in social media using deep learning.一种使用深度学习的社交媒体多模态仇恨言论检测综合框架。
Sci Rep. 2025 Apr 15;15(1):13020. doi: 10.1038/s41598-025-94069-z.
2
Brain tumor segmentation and detection in MRI using convolutional neural networks and VGG16.使用卷积神经网络和VGG16在磁共振成像(MRI)中进行脑肿瘤分割与检测
Cancer Biomark. 2025 Mar;42(3):18758592241311184. doi: 10.1177/18758592241311184. Epub 2025 Apr 4.
3
A hybrid deep learning framework for early detection of diabetic retinopathy using retinal fundus images.一种使用视网膜眼底图像进行糖尿病视网膜病变早期检测的混合深度学习框架。
Sci Rep. 2025 Apr 30;15(1):15166. doi: 10.1038/s41598-025-99309-w.
4
A novel hybrid deep learning IChOA-CNN-LSTM model for modality-enriched and multilingual emotion recognition in social media.一种新颖的混合深度学习 IChOA-CNN-LSTM 模型,用于社交媒体中丰富模态和多语言的情感识别。
Sci Rep. 2024 Sep 27;14(1):22270. doi: 10.1038/s41598-024-73452-2.
5
Detection of Hate Speech in COVID-19-Related Tweets in the Arab Region: Deep Learning and Topic Modeling Approach.检测阿拉伯地区与 COVID-19 相关推文的仇恨言论:深度学习和主题建模方法。
J Med Internet Res. 2020 Dec 8;22(12):e22609. doi: 10.2196/22609.
6
Roman Urdu Hate Speech Detection Using Transformer-Based Model for Cyber Security Applications.基于转换器模型的罗曼 Urdu 仇恨言论检测在网络安全应用中的研究
Sensors (Basel). 2023 Apr 12;23(8):3909. doi: 10.3390/s23083909.
7
Detection of hate: speech tweets based convolutional neural network and machine learning algorithms.基于卷积神经网络和机器学习算法的仇恨言论检测:推特言论。
Sci Rep. 2024 Nov 21;14(1):28870. doi: 10.1038/s41598-024-76632-2.
8
A hybrid transformer and attention based recurrent neural network for robust and interpretable sentiment analysis of tweets.基于混合变压器和注意力的循环神经网络的 tweet 情感分析的鲁棒性和可解释性。
Sci Rep. 2024 Oct 22;14(1):24882. doi: 10.1038/s41598-024-76079-5.
9
Roman urdu hate speech detection using hybrid machine learning models and hyperparameter optimization.基于混合机器学习模型和超参数优化的罗马 Urdu 仇恨言论检测
Sci Rep. 2024 Nov 19;14(1):28590. doi: 10.1038/s41598-024-79106-7.
10
Hybrid deep learning for computational precision in cardiac MRI segmentation: Integrating Autoencoders, CNNs, and RNNs for enhanced structural analysis.用于心脏磁共振成像分割计算精度的混合深度学习:整合自动编码器、卷积神经网络和循环神经网络以增强结构分析。
Comput Biol Med. 2025 Mar;186:109597. doi: 10.1016/j.compbiomed.2024.109597. Epub 2025 Jan 1.

本文引用的文献

1
A veracity dissemination consistency-based few-shot fake news detection framework by synergizing adversarial and contrastive self-supervised learning.一种基于真实性传播一致性的少样本假新闻检测框架,通过协同对抗性和对比性自监督学习实现。
Sci Rep. 2024 Aug 22;14(1):19470. doi: 10.1038/s41598-024-70039-9.
2
Transformer Architecture and Attention Mechanisms in Genome Data Analysis: A Comprehensive Review.基因组数据分析中的Transformer架构与注意力机制:全面综述
Biology (Basel). 2023 Jul 22;12(7):1033. doi: 10.3390/biology12071033.
3
The multi-modal fusion in visual question answering: a review of attention mechanisms.
视觉问答中的多模态融合:注意力机制综述
PeerJ Comput Sci. 2023 May 30;9:e1400. doi: 10.7717/peerj-cs.1400. eCollection 2023.