Jiang Le, Ma Liyan, Yang Guang
Bioengineering Department and Imperial-X, Imperial College London, London W12 7SL, UK.
School of Computer Engineering and Science, Shanghai University, Shanghai 200444, China.
Med Image Anal. 2025 Oct;105:103673. doi: 10.1016/j.media.2025.103673. Epub 2025 Jun 21.
Federated learning (FL) has emerged as a transformative framework for privacy-preserving distributed training, allowing clients to collaboratively train a global model without sharing their local data. This is especially crucial in sensitive fields like healthcare, where protecting patient data is paramount. However, privacy leakage remains a critical challenge, as the communication of model updates can be exploited by potential adversaries. Gradient inversion attacks (GIAs), for instance, allow adversaries to approximate the gradients used for training and reconstruct training images, thus stealing patient privacy. Existing defense mechanisms obscure gradients, yet lack a nuanced understanding of which gradients or types of image information are most vulnerable to such attacks. These indiscriminate calibrated perturbations result in either excessive privacy protection degrading model accuracy, or insufficient one failing to safeguard sensitive information. Therefore, we introduce a framework that addresses these challenges by leveraging a shadow model with interpretability for identifying sensitive areas. This enables a more targeted and sample-specific noise injection. Specially, our defensive strategy achieves discrepancies of 3.73 in PSNR and 0.2 in SSIM compared to the circumstance without defense on the ChestXRay dataset, and 2.78 in PSNR and 0.166 in the EyePACS dataset. Moreover, it minimizes adverse effects on model performance, with less than 1% F1 reduction compared to SOTA methods. Our extensive experiments, conducted across diverse types of medical images, validate the generalization of the proposed framework. The stable defense improvements for FedAvg are consistently over 1.5% times in LPIPS and SSIM. It also offers a universal defense against various GIA types, especially for these sensitive areas in images.
联邦学习(FL)已成为一种用于隐私保护分布式训练的变革性框架,允许客户端在不共享本地数据的情况下协作训练全局模型。这在医疗保健等敏感领域尤为关键,因为保护患者数据至关重要。然而,隐私泄露仍然是一个关键挑战,因为模型更新的通信可能会被潜在对手利用。例如,梯度反转攻击(GIA)允许对手近似用于训练的梯度并重建训练图像,从而窃取患者隐私。现有的防御机制会模糊梯度,但对哪些梯度或图像信息类型最容易受到此类攻击缺乏细致的理解。这些不加区分的校准扰动要么导致过度的隐私保护降低模型准确性,要么导致保护不足无法保护敏感信息。因此,我们引入了一个框架,通过利用具有可解释性的影子模型来识别敏感区域,从而应对这些挑战。这使得能够进行更有针对性的、针对特定样本的噪声注入。具体而言,与在ChestXRay数据集上无防御的情况相比,我们的防御策略在PSNR方面实现了3.73的差异,在SSIM方面实现了0.2的差异;在EyePACS数据集上,PSNR方面为2.78,SSIM方面为0.166。此外,它将对模型性能的不利影响降至最低,与最优方法相比,F1降低不到1%。我们在各种类型的医学图像上进行的广泛实验验证了所提出框架的通用性。对于FedAvg的稳定防御改进在LPIPS和SSIM方面始终超过1.5%。它还为各种类型的GIA提供了通用防御,特别是针对图像中的这些敏感区域。