Alghamdi Jawaher, Lin Yuqing, Luo Suhuai
School of Information and Physical Sciences, The University of Newcastle, Newcastle, Australia.
Department of Computer Science, King Khalid University, Abha, Saudi Arabia.
Knowl Based Syst. 2023 Aug 15;274:110642. doi: 10.1016/j.knosys.2023.110642. Epub 2023 May 19.
The COVID-19 pandemic has resulted in a surge of fake news, creating public health risks. However, developing an effective way to detect such news is challenging, especially when published news involves mixing true and false information. Detecting COVID-19 fake news has become a critical task in the field of natural language processing (NLP). This paper explores the effectiveness of several machine learning algorithms and fine-tuning pre-trained transformer-based models, including Bidirectional Encoder Representations from Transformers (BERT) and COVID-Twitter-BERT (CT-BERT), for COVID-19 fake news detection. We evaluate the performance of different downstream neural network structures, such as CNN and BiGRU layers, added on top of BERT and CT-BERT with frozen or unfrozen parameters. Our experiments on a real-world COVID-19 fake news dataset demonstrate that incorporating BiGRU on top of the CT-BERT model achieves outstanding performance, with a state-of-the-art F1 score of 98%. These results have significant implications for mitigating the spread of COVID-19 misinformation and highlight the potential of advanced machine learning models for fake news detection.
新冠疫情导致了虚假新闻的激增,带来了公共卫生风险。然而,开发一种有效的方法来检测此类新闻具有挑战性,尤其是当发布的新闻涉及真假信息混合时。检测新冠虚假新闻已成为自然语言处理(NLP)领域的一项关键任务。本文探讨了几种机器学习算法以及微调基于预训练Transformer的模型(包括来自Transformer的双向编码器表征(BERT)和新冠推特BERT(CT - BERT))在检测新冠虚假新闻方面的有效性。我们评估了在BERT和CT - BERT之上添加具有冻结或未冻结参数的不同下游神经网络结构(如CNN和双向门控循环单元(BiGRU)层)的性能。我们在一个真实世界的新冠虚假新闻数据集上的实验表明,在CT - BERT模型之上加入BiGRU可实现出色的性能,达到了98%的先进F1分数。这些结果对于减轻新冠错误信息的传播具有重要意义,并凸显了先进机器学习模型在虚假新闻检测方面的潜力。