Department of Computer Science, Bahria University, Karachi, Pakistan.
PLoS One. 2024 Oct 21;19(10):e0307363. doi: 10.1371/journal.pone.0307363. eCollection 2024.
Convolutional Neural Network (CNN)-based models are prone to adversarial attacks, which present a significant hurdle to their reliability and robustness. The vulnerability of CNN-based models may be exploited by attackers to launch cyber-attacks. An attacker typically adds small, carefully crafted perturbations to original medical images. When a CNN-based model receives the perturbed medical image as input, it misclassifies the image, even though the added perturbation is often imperceptible to the human eye. The emergence of such attacks has raised security concerns regarding the implementation of deep learning-based medical image classification systems within clinical environments. To address this issue, a reliable defense mechanism is required to detect adversarial attacks on medical images. This study will focus on the robust detection of pneumonia in chest X-ray images through CNN-based models. Various adversarial attacks and defense strategies will be evaluated and analyzed in the context of CNN-based pneumonia detection. From earlier studies, it has been observed that a single defense mechanism is usually not effective against more than one type of adversarial attack. Therefore, this study will propose a defense mechanism that is effective against multiple attack types. A reliable defense framework for pneumonia detection models will ensure secure clinical deployment, facilitating radiologists and doctors in their diagnosis and treatment planning. It can also save time and money by automating routine tasks. The proposed defense mechanism includes a convolutional autoencoder to denoise perturbed Fast Gradient Sign Method (FGSM) and Projected Gradient Descent (PGD) adversarial images, two state- of-the-art attacks carried out at five magnitudes, i.e., ε (epsilon) values. Two pre-trained models, VGG19 and VGG16, and our hybrid model of MobileNetV2 and DenseNet169, called Stack Model, have been used to compare their results. This study shows that the proposed defense mechanism outperforms state-of-the-art studies. The PGD attack using the VGG16 model shows a better attack success rate by reducing overall accuracy by up to 67%. The autoencoder improves accuracy by up to 16% against PGD attacks in both the VGG16 and VGG19 models.
基于卷积神经网络(CNN)的模型容易受到对抗攻击的影响,这对它们的可靠性和稳健性构成了重大障碍。攻击者可能会利用基于 CNN 的模型的漏洞发起网络攻击。攻击者通常会向原始医学图像添加小的、精心设计的扰动。当基于 CNN 的模型接收被扰动的医学图像作为输入时,它会错误分类图像,尽管添加的扰动通常人眼无法察觉。这些攻击的出现引起了人们对深度学习为基础的医学图像分类系统在临床环境中的实施的安全问题的关注。为了解决这个问题,需要一个可靠的防御机制来检测医学图像上的对抗攻击。本研究将专注于通过基于 CNN 的模型来稳健地检测胸部 X 射线图像中的肺炎。将在基于 CNN 的肺炎检测的背景下评估和分析各种对抗攻击和防御策略。从早期的研究中可以看出,单一的防御机制通常不能有效地抵御多种类型的对抗攻击。因此,本研究将提出一种针对多种攻击类型有效的防御机制。为肺炎检测模型提供可靠的防御框架将确保安全的临床部署,为放射科医生和医生提供诊断和治疗计划的便利。它还可以通过自动化常规任务来节省时间和金钱。所提出的防御机制包括一个卷积自动编码器,用于对扰动的快速梯度符号法(Fast Gradient Sign Method,FGSM)和投影梯度下降(Projected Gradient Descent,PGD)对抗图像进行去噪,这两种是在五个幅度上进行的最先进的攻击,即ε(epsilon)值。使用了两个预训练的模型,VGG19 和 VGG16,以及我们的 MobileNetV2 和 DenseNet169 的混合模型,称为 Stack Model,用于比较它们的结果。本研究表明,所提出的防御机制优于最先进的研究。使用 VGG16 模型的 PGD 攻击通过将整体准确率降低高达 67%,显示出更好的攻击成功率。在 VGG16 和 VGG19 模型中,自动编码器针对 PGD 攻击将准确率提高了高达 16%。