Minagi Akinori, Hirano Hokuto, Takemoto Kauzhiro
Department of Bioscience and Bioinformatics, Kyushu Institute of Technology, Iizuka 820-8502, Fukuoka, Japan.
J Imaging. 2022 Feb 4;8(2):38. doi: 10.3390/jimaging8020038.
Transfer learning from natural images is used in deep neural networks (DNNs) for medical image classification to achieve a computer-aided clinical diagnosis. Although the adversarial vulnerability of DNNs hinders practical applications owing to the high stakes of diagnosis, adversarial attacks are expected to be limited because training datasets (medical images), which are often required for adversarial attacks, are generally unavailable in terms of security and privacy preservation. Nevertheless, in this study, we demonstrated that adversarial attacks are also possible using natural images for medical DNN models with transfer learning, even if such medical images are unavailable; in particular, we showed that universal adversarial perturbations (UAPs) can also be generated from natural images. UAPs from natural images are useful for both non-targeted and targeted attacks. The performance of UAPs from natural images was significantly higher than that of random controls. The use of transfer learning causes a security hole, which decreases the reliability and safety of computer-based disease diagnosis. Model training from random initialization reduced the performance of UAPs from natural images; however, it did not completely avoid vulnerability to UAPs. The vulnerability of UAPs to natural images is expected to become a significant security threat.
从自然图像进行迁移学习被应用于深度神经网络(DNN)中,用于医学图像分类以实现计算机辅助临床诊断。尽管由于诊断的高风险,DNN的对抗性脆弱性阻碍了其实际应用,但由于对抗性攻击通常所需的训练数据集(医学图像)在安全和隐私保护方面通常不可用,预计对抗性攻击会受到限制。然而,在本研究中,我们证明了即使没有医学图像,使用自然图像对具有迁移学习的医学DNN模型进行对抗性攻击也是可能的;特别是,我们表明也可以从自然图像生成通用对抗扰动(UAP)。来自自然图像的UAP对非目标攻击和目标攻击都很有用。来自自然图像的UAP的性能显著高于随机对照。迁移学习的使用导致了一个安全漏洞,这降低了基于计算机的疾病诊断的可靠性和安全性。从随机初始化进行模型训练降低了来自自然图像的UAP的性能;然而,它并没有完全避免对UAP的脆弱性。UAP对自然图像的脆弱性预计将成为一个重大的安全威胁。