Digital Engineering Faculty, University of Potsdam, 14482 Potsdam, Germany.
Hasso Plattner Institute for Digital Engineering gGmbH, 14482 Potsdam, Germany.
Sensors (Basel). 2022 Jul 11;22(14):5195. doi: 10.3390/s22145195.
Privacy regulations and the physical distribution of heterogeneous data are often primary concerns for the development of deep learning models in a medical context. This paper evaluates the feasibility of differentially private federated learning for chest X-ray classification as a defense against data privacy attacks. To the best of our knowledge, we are the first to directly compare the impact of differentially private training on two different neural network architectures, DenseNet121 and ResNet50. Extending the federated learning environments previously analyzed in terms of privacy, we simulated a heterogeneous and imbalanced federated setting by distributing images from the public CheXpert and Mendeley chest X-ray datasets unevenly among 36 clients. Both non-private baseline models achieved an area under the receiver operating characteristic curve (AUC) of 0.94 on the binary classification task of detecting the presence of a medical finding. We demonstrate that both model architectures are vulnerable to privacy violation by applying image reconstruction attacks to local model updates from individual clients. The attack was particularly successful during later training stages. To mitigate the risk of a privacy breach, we integrated Rényi differential privacy with a Gaussian noise mechanism into local model training. We evaluate model performance and attack vulnerability for privacy budgets ε∈{1,3,6,10}. The DenseNet121 achieved the best utility-privacy trade-off with an AUC of 0.94 for ε=6. Model performance deteriorated slightly for individual clients compared to the non-private baseline. The ResNet50 only reached an AUC of 0.76 in the same privacy setting. Its performance was inferior to that of the DenseNet121 for all considered privacy constraints, suggesting that the DenseNet121 architecture is more robust to differentially private training.
隐私法规和异质数据的物理分布通常是医学背景下深度学习模型发展的主要关注点。本文评估了差分隐私联邦学习在胸部 X 射线分类中的可行性,作为抵御数据隐私攻击的一种防御措施。据我们所知,我们是第一个直接比较差分隐私训练对两种不同神经网络架构(DenseNet121 和 ResNet50)影响的人。在隐私方面扩展之前分析的联邦学习环境,我们通过在 36 个客户端之间不均匀地分发来自公共 CheXpert 和 Mendeley 胸部 X 射线数据集的图像,模拟了一个异构和不平衡的联邦设置。两个非私有基线模型在检测医学发现存在的二进制分类任务上都达到了 0.94 的接收器操作特征曲线(AUC)下面积。我们通过对来自各个客户端的本地模型更新应用图像重建攻击,证明了这两个模型架构都容易受到隐私侵犯。攻击在后期训练阶段尤其成功。为了降低隐私泄露的风险,我们将 Rényi 差分隐私与高斯噪声机制集成到本地模型训练中。我们评估了隐私预算 ε∈{1,3,6,10}下的模型性能和攻击脆弱性。DenseNet121 在 ε=6 时达到了最佳的效用-隐私权衡,AUC 为 0.94。与非私有基线相比,个别客户端的模型性能略有下降。ResNet50 在相同的隐私设置下仅达到 AUC 为 0.76。在所有考虑的隐私约束下,它的性能都不如 DenseNet121,这表明 DenseNet121 架构对差分隐私训练更具鲁棒性。