Alabrah Amerah
Department of Information Systems, College of Computer and Information Sciences, King Saud University, Riyadh, Saudi Arabia.
PeerJ Comput Sci. 2025 Apr 16;11:e2803. doi: 10.7717/peerj-cs.2803. eCollection 2025.
Image forgery is an increasing threat, fueling misinformation and potentially impacting legal decisions and everyday life. Detecting forged media, including images and videos, is crucial for preserving trust and integrity across various platforms. Common forgery techniques like copy-move and splicing require robust detection methods to identify tampered areas without explicit guidance. The previously proposed studies focused on a single type of forgery detection utilizing block-based and key-point feature selection-based classical machine learning (ML) approaches. Furthermore, applied deep learning (DL) methods only focus on deep feature extraction without considering the focus on tampered regions detection or any domain-specific loss. Therefore, this study addresses the aforementioned challenges by proposing a lightweight DL approach, a self-supervised, triplet and auxiliary losses-based forgery detection network (SFTA-Net), featuring a self-guidance mechanism for detecting tampered regions with a commutative loss within images. The SFTA-Net method is proposed to classify forged and original photos belonging to copy-move and splicing forgeries. To effectively analyze the added components in the proposed model, three experiments were conducted, one with a self-guided (SG) head-based convolutional neural network (CNN), a second with SG-head and auxiliary loss, and a third one with SG-head auxiliary loss and triplet losses-based CNN. For experimentation, CASIA 1.0 and CASIA 2.0 datasets were used with 80-10-10% train-validation and test ratios. The testing results achieved on CASIA 1.0 were 95% accuracy and 97% accuracy on the CASIA 2.0 dataset. To prove the approach's robustness and generalization, the CASIA 2.0-trained weights were used to test on the MICC-FC2000 dataset and yielded limited results. To improve the results, fine-tuning was performed on CASIA 2.0 weights utilizing the MICC-FC2000 dataset which achieved 98% accurate results. Our findings demonstrate that the SFTA-Net surpasses the baseline ResNet18 model and previous state-of-the-art (SOTA) methods. Overall, our SG approach offers a promising solution for detecting forged images across diverse real-world scenarios, contributing to the mitigation of image forgery and preservation of trust in digital content.
图像伪造构成的威胁日益增大,助长了错误信息的传播,并可能影响法律判决和日常生活。检测包括图像和视频在内的伪造媒体对于维护各个平台的信任和完整性至关重要。诸如复制-粘贴和拼接等常见的伪造技术需要强大的检测方法来识别被篡改区域,而无需明确的指导。先前提出的研究集中在利用基于块和基于关键点特征选择的经典机器学习(ML)方法进行单一类型的伪造检测。此外,应用的深度学习(DL)方法仅专注于深度特征提取,而没有考虑对被篡改区域检测的关注或任何特定领域的损失。因此,本研究通过提出一种轻量级的DL方法来应对上述挑战,即一种基于自监督、三元组和辅助损失的伪造检测网络(SFTA-Net),其具有一种自引导机制,用于通过图像内的可交换损失来检测被篡改区域。提出SFTA-Net方法来对属于复制-粘贴和拼接伪造的伪造照片和原始照片进行分类。为了有效地分析所提出模型中添加的组件,进行了三个实验,一个实验使用基于自引导(SG)头的卷积神经网络(CNN),第二个实验使用SG头和辅助损失,第三个实验使用基于SG头辅助损失和三元组损失的CNN。为了进行实验,使用了CASIA 1.0和CASIA 2.0数据集,训练-验证和测试比例为80-10-10%。在CASIA 1.0上取得的测试结果准确率为95%,在CASIA 2.0数据集上的准确率为97%。为了证明该方法的鲁棒性和泛化能力,使用在CASIA 2.0上训练的权重在MICC-FC2000数据集上进行测试,结果有限。为了改进结果,利用MICC-FC2000数据集对CASIA 2.0权重进行微调,取得了98%的准确结果。我们的研究结果表明,SFTA-Net优于基线ResNet18模型和先前的最先进(SOTA)方法。总体而言,我们的SG方法为跨各种现实世界场景检测伪造图像提供了一个有前景的解决方案,有助于减轻图像伪造并维护对数字内容的信任。