Department of Health Outcomes and Biomedical Informatics, University of Florida, Gainesville, FL, USA.
Pharmaceutical Outcomes & Policy, University of Florida, Gainesville, FL, USA.
J Biomed Inform. 2024 Mar;151:104622. doi: 10.1016/j.jbi.2024.104622. Epub 2024 Mar 6.
The integration of artificial intelligence (AI) and machine learning (ML) in health care to aid clinical decisions is widespread. However, as AI and ML take important roles in health care, there are concerns about AI and ML associated fairness and bias. That is, an AI tool may have a disparate impact, with its benefits and drawbacks unevenly distributed across societal strata and subpopulations, potentially exacerbating existing health inequities. Thus, the objectives of this scoping review were to summarize existing literature and identify gaps in the topic of tackling algorithmic bias and optimizing fairness in AI/ML models using real-world data (RWD) in health care domains.
We conducted a thorough review of techniques for assessing and optimizing AI/ML model fairness in health care when using RWD in health care domains. The focus lies on appraising different quantification metrics for accessing fairness, publicly accessible datasets for ML fairness research, and bias mitigation approaches.
We identified 11 papers that are focused on optimizing model fairness in health care applications. The current research on mitigating bias issues in RWD is limited, both in terms of disease variety and health care applications, as well as the accessibility of public datasets for ML fairness research. Existing studies often indicate positive outcomes when using pre-processing techniques to address algorithmic bias. There remain unresolved questions within the field that require further research, which includes pinpointing the root causes of bias in ML models, broadening fairness research in AI/ML with the use of RWD and exploring its implications in healthcare settings, and evaluating and addressing bias in multi-modal data.
This paper provides useful reference material and insights to researchers regarding AI/ML fairness in real-world health care data and reveals the gaps in the field. Fair AI/ML in health care is a burgeoning field that requires a heightened research focus to cover diverse applications and different types of RWD.
人工智能(AI)和机器学习(ML)在医疗保健中的融合,有助于辅助临床决策,这一应用已十分广泛。然而,随着 AI 和 ML 在医疗保健领域发挥着重要作用,人们开始关注 AI 和 ML 相关的公平性和偏见问题。也就是说,AI 工具可能会产生不同的影响,其优势和劣势在社会阶层和亚人群中分布不均,从而可能加剧现有的健康不平等现象。因此,本研究的目的是总结现有文献,并确定在使用真实世界数据(RWD)解决 AI/ML 模型中的算法偏差和优化公平性的主题方面存在的差距。
我们对使用 RWD 在医疗保健领域中评估和优化 AI/ML 模型公平性的技术进行了全面审查。重点在于评估用于评估公平性的不同量化指标、用于 ML 公平性研究的公共数据集以及偏差缓解方法。
我们确定了 11 篇专注于优化医疗保健应用中模型公平性的论文。目前关于减轻 RWD 中偏差问题的研究在疾病种类和医疗保健应用方面都很有限,同时用于 ML 公平性研究的公共数据集也很有限。现有研究表明,在使用预处理技术解决算法偏差问题时,通常会产生积极的结果。该领域仍存在一些悬而未决的问题,需要进一步研究,包括确定 ML 模型中偏差的根本原因、使用 RWD 拓宽 AI/ML 中的公平性研究并探索其在医疗保健环境中的影响,以及评估和解决多模态数据中的偏差问题。
本文为研究人员提供了有关真实世界医疗保健数据中 AI/ML 公平性的有用参考材料和见解,并揭示了该领域的差距。医疗保健中的公平 AI/ML 是一个新兴领域,需要更加关注不同的应用和不同类型的 RWD,以覆盖更广泛的领域。