Department of Bioscience and Bioinformatics, Kyushu Institute of Technology, Iizuka, Fukuoka, 820-8502, Japan.
BMC Med Imaging. 2021 Jan 7;21(1):9. doi: 10.1186/s12880-020-00530-y.
Deep neural networks (DNNs) are widely investigated in medical image classification to achieve automated support for clinical diagnosis. It is necessary to evaluate the robustness of medical DNN tasks against adversarial attacks, as high-stake decision-making will be made based on the diagnosis. Several previous studies have considered simple adversarial attacks. However, the vulnerability of DNNs to more realistic and higher risk attacks, such as universal adversarial perturbation (UAP), which is a single perturbation that can induce DNN failure in most classification tasks has not been evaluated yet.
We focus on three representative DNN-based medical image classification tasks (i.e., skin cancer, referable diabetic retinopathy, and pneumonia classifications) and investigate their vulnerability to the seven model architectures of UAPs.
We demonstrate that DNNs are vulnerable to both nontargeted UAPs, which cause a task failure resulting in an input being assigned an incorrect class, and to targeted UAPs, which cause the DNN to classify an input into a specific class. The almost imperceptible UAPs achieved > 80% success rates for nontargeted and targeted attacks. The vulnerability to UAPs depended very little on the model architecture. Moreover, we discovered that adversarial retraining, which is known to be an effective method for adversarial defenses, increased DNNs' robustness against UAPs in only very few cases.
Unlike previous assumptions, the results indicate that DNN-based clinical diagnosis is easier to deceive because of adversarial attacks. Adversaries can cause failed diagnoses at lower costs (e.g., without consideration of data distribution); moreover, they can affect the diagnosis. The effects of adversarial defenses may not be limited. Our findings emphasize that more careful consideration is required in developing DNNs for medical imaging and their practical applications.
深度神经网络(DNN)在医学图像分类中得到了广泛的研究,以实现对临床诊断的自动化支持。有必要评估医学 DNN 任务对对抗攻击的稳健性,因为基于诊断的决策将具有高风险。之前的一些研究已经考虑了简单的对抗攻击。然而,DNN 对更现实和更高风险的攻击(例如通用对抗扰动(UAP))的脆弱性尚未得到评估,UAP 是一种可以在大多数分类任务中导致 DNN 故障的单一扰动。
我们专注于三个代表性的基于 DNN 的医学图像分类任务(即皮肤癌、可转诊糖尿病视网膜病变和肺炎分类),并研究它们对 UAP 的七个模型架构的脆弱性。
我们证明了 DNN 容易受到非目标 UAP 的影响,非目标 UAP 会导致任务失败,导致输入被分配到错误的类别,以及目标 UAP 的影响,目标 UAP 会导致 DNN 将输入分类为特定类别。几乎不可察觉的 UAP 针对非目标和目标攻击实现了超过 80%的成功率。对 UAP 的脆弱性几乎与模型架构无关。此外,我们发现对抗性再训练,这是一种对抗防御的有效方法,在极少数情况下增加了 DNN 对 UAP 的鲁棒性。
与之前的假设不同,结果表明,基于 DNN 的临床诊断更容易受到对抗攻击的欺骗。攻击者可以以较低的成本(例如,不考虑数据分布)导致诊断失败;此外,它们会影响诊断。对抗防御的效果可能不会受到限制。我们的研究结果强调,在开发用于医学成像和实际应用的 DNN 时,需要更加谨慎。