Kumar K Naveen, Mohan C Krishna, Cenkeramaddi Linga Reddy, Awasthi Navchetan
Department of Computer Science and Engineering, Indian Institute of Technology Hyderabad (IITH), Hyderabad, 502284, India.
Department of Information and Communication Technology, University of Agder, Grimstad, Norway.
Artif Intell Med. 2025 Jan;159:103024. doi: 10.1016/j.artmed.2024.103024. Epub 2024 Nov 26.
The privacy-sensitive nature of medical image data is often bounded by strict data sharing regulations that necessitate the need for novel modeling and analysis techniques. Federated learning (FL) enables multiple medical institutions to collectively train a deep neural network without sharing sensitive patient information. In addition, FL uses its collaborative approach to address challenges related to the scarcity and non-uniform distribution of heterogeneous medical domain data. Nevertheless, the data-opaque nature and distributed setup make FL susceptible to data poisoning attacks. There are diverse FL data poisoning attacks for classification models on natural image data in the literature. But their primary focus is on the impact of the attack and they do not consider the attack budget and attack visibility. The attack budget is essential for adversaries to optimize resource utilization in real-world scenarios, which determines the number of manipulations or perturbations they can apply. Simultaneously, attack visibility is crucial to ensure covert execution, allowing attackers to achieve their objectives without triggering detection mechanisms. Generally, an attacker's aim is to create maximum attack impact with minimal resources and low visibility. So, considering these three entities can effectively comprehend the adversary's perspective in designing an attack for real-world scenarios. Further, data poisoning attacks on medical images are challenging compared to natural images due to the subjective nature of medical data. Hence, we develop an attack with a low budget, low visibility, and high impact for medical image classification in FL. We propose a federated learning attention guided minimal attack (FL-AGMA), that uses class attention maps to identify specific medical image regions for perturbation. We introduce image distortion degree (IDD) as a metric to assess the attack budget. Also, we develop a feedback mechanism to regulate the attack coefficient for low attack visibility. Later, we optimize the attack budget by adaptively changing the IDD based on attack visibility. We extensively evaluate three large-scale datasets, namely, Covid-chestxray, Camelyon17, and HAM10000, covering three different data modalities. We observe that our FL-AGMA method has resulted in 44.49% less test accuracy with only 24% of IDD attack budget and lower attack visibility compared to the other attacks.
医学图像数据的隐私敏感性往往受到严格的数据共享法规的限制,这就需要新颖的建模和分析技术。联邦学习(FL)使多个医疗机构能够在不共享敏感患者信息的情况下共同训练深度神经网络。此外,联邦学习采用协作方法来应对异构医学领域数据稀缺和分布不均相关的挑战。然而,数据不透明的性质和分布式设置使联邦学习容易受到数据中毒攻击。文献中有针对自然图像数据分类模型的各种联邦学习数据中毒攻击。但它们主要关注攻击的影响,没有考虑攻击预算和攻击可见性。攻击预算对于对手在现实场景中优化资源利用至关重要,它决定了他们可以应用的操作或扰动的数量。同时,攻击可见性对于确保隐蔽执行至关重要,使攻击者能够在不触发检测机制的情况下实现其目标。一般来说,攻击者的目标是以最少的资源和低可见性创造最大的攻击影响。因此,考虑这三个因素可以有效地从对手的角度理解为现实场景设计攻击。此外,由于医学数据的主观性,对医学图像的数据中毒攻击比自然图像更具挑战性。因此,我们针对联邦学习中的医学图像分类开发了一种低预算、低可见性和高影响的攻击方法。我们提出了一种联邦学习注意力引导最小攻击(FL-AGMA),它使用类别注意力图来识别特定的医学图像区域进行扰动。我们引入图像失真度(IDD)作为评估攻击预算的指标。此外,我们开发了一种反馈机制来调节攻击系数以实现低攻击可见性。之后,我们根据攻击可见性自适应地改变IDD来优化攻击预算。我们广泛评估了三个大规模数据集,即Covid-chestxray、Camelyon17和HAM10000,涵盖三种不同的数据模态。我们观察到,与其他攻击相比,我们的FL-AGMA方法在仅24%的IDD攻击预算和更低的攻击可见性情况下,测试准确率降低了44.49%。