• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于损害分类的推文:使用微调BERT模型进行基础设施和人员损害评估

Categorization of tweets for damages: infrastructure and human damage assessment using fine-tuned BERT model.

作者信息

Malik Muhammad Shahid Iqbal, Younas Muhammad Zeeshan, Jamjoom Mona Mamdouh, Ignatov Dmitry I

机构信息

Department of Computer Science, National Research University Higher School of Economics, Moscow, Russia.

Department of Computer Science, Capital University of Science and Technology, Islamabad, Pakistan.

出版信息

PeerJ Comput Sci. 2024 Feb 16;10:e1859. doi: 10.7717/peerj-cs.1859. eCollection 2024.

DOI:10.7717/peerj-cs.1859
PMID:38435619
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10909225/
Abstract

Identification of infrastructure and human damage assessment tweets is beneficial to disaster management organizations as well as victims during a disaster. Most of the prior works focused on the detection of informative/situational tweets, and infrastructure damage, only one focused on human damage. This study presents a novel approach for detecting damage assessment tweets involving infrastructure and human damages. We investigated the potential of the Bidirectional Encoder Representations from Transformer (BERT) model to learn universal contextualized representations targeting to demonstrate its effectiveness for binary and multi-class classification of disaster damage assessment tweets. The objective is to exploit a pre-trained BERT as a transfer learning mechanism after fine-tuning important hyper-parameters on the CrisisMMD dataset containing seven disasters. The effectiveness of fine-tuned BERT is compared with five benchmarks and nine comparable models by conducting exhaustive experiments. The findings show that the fine-tuned BERT outperformed all benchmarks and comparable models and achieved state-of-the-art performance by demonstrating up to 95.12% macro-f1-score, and 88% macro-f1-score for binary and multi-class classification. Specifically, the improvement in the classification of human damage is promising.

摘要

识别有关基础设施和人员损害评估的推文,对灾害管理组织以及灾难中的受害者都有益处。大多数先前的工作都集中在检测信息性/情境性推文以及基础设施损害方面,只有一项研究关注人员损害。本研究提出了一种新颖的方法来检测涉及基础设施和人员损害的损害评估推文。我们研究了来自Transformer的双向编码器表征(BERT)模型学习通用上下文表征的潜力,旨在证明其在灾害损害评估推文的二分类和多分类中的有效性。目标是在包含七种灾害的CrisisMMD数据集上微调重要超参数后,将预训练的BERT用作迁移学习机制。通过进行详尽的实验,将微调后的BERT的有效性与五个基准和九个可比模型进行了比较。研究结果表明,微调后的BERT优于所有基准和可比模型,并通过在二分类和多分类中分别展示高达95.12%的宏F1分数和88%的宏F1分数,取得了当前最优的性能。具体而言,在人员损害分类方面的改进很有前景。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fa6b/10909225/19d0f1bfc0e0/peerj-cs-10-1859-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fa6b/10909225/d27775372b55/peerj-cs-10-1859-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fa6b/10909225/a3514c401a08/peerj-cs-10-1859-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fa6b/10909225/01c644406640/peerj-cs-10-1859-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fa6b/10909225/6904cb69a289/peerj-cs-10-1859-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fa6b/10909225/19d0f1bfc0e0/peerj-cs-10-1859-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fa6b/10909225/d27775372b55/peerj-cs-10-1859-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fa6b/10909225/a3514c401a08/peerj-cs-10-1859-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fa6b/10909225/01c644406640/peerj-cs-10-1859-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fa6b/10909225/6904cb69a289/peerj-cs-10-1859-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fa6b/10909225/19d0f1bfc0e0/peerj-cs-10-1859-g005.jpg

相似文献

1
Categorization of tweets for damages: infrastructure and human damage assessment using fine-tuned BERT model.用于损害分类的推文:使用微调BERT模型进行基础设施和人员损害评估
PeerJ Comput Sci. 2024 Feb 16;10:e1859. doi: 10.7717/peerj-cs.1859. eCollection 2024.
2
Fine-Tuning BERT Models to Classify Misinformation on Garlic and COVID-19 on Twitter.微调 BERT 模型以在 Twitter 上对大蒜和 COVID-19 相关的错误信息进行分类。
Int J Environ Res Public Health. 2022 Apr 22;19(9):5126. doi: 10.3390/ijerph19095126.
3
Traditional Machine Learning Models and Bidirectional Encoder Representations From Transformer (BERT)-Based Automatic Classification of Tweets About Eating Disorders: Algorithm Development and Validation Study.传统机器学习模型与基于双向编码器表征变换器(BERT)的饮食失调推文自动分类:算法开发与验证研究
JMIR Med Inform. 2022 Feb 24;10(2):e34492. doi: 10.2196/34492.
4
An efficient method for disaster tweets classification using gradient-based optimized convolutional neural networks with BERT embeddings.一种使用基于梯度优化的卷积神经网络与BERT嵌入的高效灾难推文分类方法。
MethodsX. 2024 Jul 3;13:102843. doi: 10.1016/j.mex.2024.102843. eCollection 2024 Dec.
5
Comparing Pre-trained and Feature-Based Models for Prediction of Alzheimer's Disease Based on Speech.基于语音比较预训练模型和基于特征的模型对阿尔茨海默病的预测
Front Aging Neurosci. 2021 Apr 27;13:635945. doi: 10.3389/fnagi.2021.635945. eCollection 2021.
6
Comparison of pretrained transformer-based models for influenza and COVID-19 detection using social media text data in Saskatchewan, Canada.加拿大萨斯喀彻温省使用社交媒体文本数据对基于预训练变压器的流感和新冠病毒检测模型的比较
Front Digit Health. 2023 Jun 28;5:1203874. doi: 10.3389/fdgth.2023.1203874. eCollection 2023.
7
Automatic text classification of actionable radiology reports of tinnitus patients using bidirectional encoder representations from transformer (BERT) and in-domain pre-training (IDPT).使用基于转换器的双向编码器表示 (BERT) 和领域内预训练 (IDPT) 对耳鸣患者的可操作放射学报告进行自动文本分类。
BMC Med Inform Decis Mak. 2022 Jul 30;22(1):200. doi: 10.1186/s12911-022-01946-y.
8
CACER: Clinical concept Annotations for Cancer Events and Relations.CACER:癌症事件与关系的临床概念注释。
J Am Med Inform Assoc. 2024 Nov 1;31(11):2583-2594. doi: 10.1093/jamia/ocae231.
9
Detecting Potentially Harmful and Protective Suicide-Related Content on Twitter: Machine Learning Approach.在 Twitter 上检测潜在有害和保护自杀相关内容:机器学习方法。
J Med Internet Res. 2022 Aug 17;24(8):e34705. doi: 10.2196/34705.
10
Fine-Tuning Large Language Models to Enhance Programmatic Assessment in Graduate Medical Education.微调大语言模型以加强毕业后医学教育中的程序化评估。
J Educ Perioper Med. 2024 Sep 30;26(3):E729. doi: 10.46374/VolXXVI_Issue3_Moore. eCollection 2024 Jul-Sep.

引用本文的文献

1
MSBKA: A Multi-Strategy Improved Black-Winged Kite Algorithm for Feature Selection of Natural Disaster Tweets Classification.MSBKA:一种用于自然灾害推文分类特征选择的多策略改进黑翅鸢算法
Biomimetics (Basel). 2025 Jan 10;10(1):41. doi: 10.3390/biomimetics10010041.

本文引用的文献

1
How to detect propaganda from social media? Exploitation of semantic and fine-tuned language models.如何从社交媒体中检测宣传内容?语义和微调语言模型的应用
PeerJ Comput Sci. 2023 Feb 20;9:e1248. doi: 10.7717/peerj-cs.1248. eCollection 2023.
2
Identification of offensive language in Urdu using semantic and embedding models.使用语义和嵌入模型识别乌尔都语中的冒犯性语言。
PeerJ Comput Sci. 2022 Dec 12;8:e1169. doi: 10.7717/peerj-cs.1169. eCollection 2022.
3
Rumour identification on Twitter as a function of novel textual and language-context features.
基于新颖文本和语言上下文特征的推特谣言识别
Multimed Tools Appl. 2023;82(5):7017-7038. doi: 10.1007/s11042-022-13595-4. Epub 2022 Aug 12.