Poulain Raphael, Tarek Mirza Farhan Bin, Beheshti Rahmatollah
University of Delaware USA.
FAccT 23 (2023). 2023 Jun;2023:1599-1608. doi: 10.1145/3593013.3594102. Epub 2023 Jun 12.
Developing AI tools that preserve fairness is of critical importance, specifically in high-stakes applications such as those in healthcare. However, health AI models' overall prediction performance is often prioritized over the possible biases such models could have. In this study, we show one possible approach to mitigate bias concerns by having healthcare institutions collaborate through a federated learning paradigm (FL; which is a popular choice in healthcare settings). While FL methods with an emphasis on fairness have been previously proposed, their underlying model and local implementation techniques, as well as their possible applications to the healthcare domain remain widely underinvestigated. Therefore, we propose a comprehensive FL approach with adversarial debiasing and a fair aggregation method, suitable to various fairness metrics, in the healthcare domain where electronic health records are used. Not only our approach explicitly mitigates bias as part of the optimization process, but an FL-based paradigm would also implicitly help with addressing data imbalance and increasing the data size, offering a practical solution for healthcare applications. We empirically demonstrate our method's superior performance on multiple experiments simulating large-scale real-world scenarios and compare it to several baselines. Our method has achieved promising fairness performance with the lowest impact on overall discrimination performance (accuracy). Our code is available at https://github.com/healthylaife/FairFedAvg.
开发确保公平性的人工智能工具至关重要,特别是在医疗保健等高风险应用领域。然而,健康人工智能模型的整体预测性能往往比此类模型可能存在的偏差更受重视。在本研究中,我们展示了一种可能的方法,即通过让医疗机构通过联邦学习范式(FL;这是医疗保健环境中的一种常用选择)进行协作,来减轻对偏差的担忧。虽然此前已经提出了强调公平性的FL方法,但其基础模型和本地实施技术,以及它们在医疗保健领域的可能应用仍未得到充分研究。因此,我们提出了一种全面的FL方法,该方法具有对抗性去偏和公平聚合方法,适用于使用电子健康记录的医疗保健领域中的各种公平性指标。我们的方法不仅在优化过程中明确减轻了偏差,而且基于FL的范式还将在隐式层面上有助于解决数据不平衡问题并增加数据量,为医疗保健应用提供了切实可行的解决方案。我们通过模拟大规模现实场景的多项实验,实证证明了我们方法的卓越性能,并将其与多个基线进行了比较。我们的方法在对整体歧视性能(准确性)影响最小的情况下,实现了令人满意的公平性表现。我们的代码可在https://github.com/healthylaife/FairFedAvg获取。