IEEE Trans Cybern. 2023 Aug;53(8):5323-5335. doi: 10.1109/TCYB.2022.3209175. Epub 2023 Jul 18.
Deep neural network has shown a powerful performance in the medical image analysis of a variety of diseases. However, a number of studies over the past few years have demonstrated that these deep learning systems can be vulnerable to well-designed adversarial attacks, with minor disruptions added to the input. Since both the public and academia have focused on deep learning in the health information economy, these adversarial attacks would prove more important and raise security concerns. In this article, adversarial attacks on deep learning systems in medicine are analyzed from two different points of view: 1) white box and 2) black box. A fast adversarial sample generation method, Feature Space-Restricted Attention Attack is proposed to explore more confusing adversarial samples. It is based on a generative adversarial network with bound classification space to generate perturbations to achieve attacks. Meanwhile, it can employ an attention mechanism to focus this perturbation on the lesion region. This enables the perturbation closely associated with the classification information making the attack more efficient and invisible. The performance and specificity of the proposed attack method are demonstrated by conducting extensive experiments on three different types of medical images. Finally, it is expected that this work can assist practitioners become being of current weaknesses in the deployment of deep learning systems in clinical settings. And, it further investigates domain-specific features of medical deep learning systems to enhance model generalization and resistance to attacks.
深度神经网络在各种疾病的医学图像分析中表现出了强大的性能。然而,过去几年的许多研究表明,这些深度学习系统容易受到精心设计的对抗攻击的影响,只需在输入中添加微小的干扰。由于公众和学术界都专注于健康信息经济中的深度学习,这些对抗攻击将变得更加重要,并引发安全担忧。本文从两个不同的角度分析了医学中深度学习系统的对抗攻击:1)白盒和 2)黑盒。提出了一种快速对抗样本生成方法,即特征空间受限注意攻击,以探索更具迷惑性的对抗样本。它基于具有受限分类空间的生成对抗网络生成扰动以实现攻击。同时,它可以采用注意力机制将这种扰动集中在病变区域。这使得扰动与分类信息紧密相关,从而使攻击更加高效和不可见。通过在三种不同类型的医学图像上进行广泛的实验,验证了所提出的攻击方法的性能和特异性。最后,希望这项工作能够帮助从业者了解在临床环境中部署深度学习系统的当前弱点,并进一步研究医学深度学习系统的特定领域特征,以提高模型泛化能力和对抗攻击的能力。