Suppr超能文献

调查医疗保健算法中的偏差:对肝脏疾病预测中监督机器学习模型的性别分层分析。

Investigating for bias in healthcare algorithms: a sex-stratified analysis of supervised machine learning models in liver disease prediction.

机构信息

Institute of Health Informatics, University College London, London, UK

Institute of Health Informatics, University College London, London, UK.

出版信息

BMJ Health Care Inform. 2022 Apr;29(1). doi: 10.1136/bmjhci-2021-100457.

Abstract

OBJECTIVES

The Indian Liver Patient Dataset (ILPD) is used extensively to create algorithms that predict liver disease. Given the existing research describing demographic inequities in liver disease diagnosis and management, these algorithms require scrutiny for potential biases. We address this overlooked issue by investigating ILPD models for sex bias.

METHODS

Following our literature review of ILPD papers, the models reported in existing studies are recreated and then interrogated for bias. We define four experiments, training on sex-unbalanced/balanced data, with and without feature selection. We build random forests (RFs), support vector machines (SVMs), Gaussian Naïve Bayes and logistic regression (LR) classifiers, running experiments 100 times, reporting average results with SD.

RESULTS

We reproduce published models achieving accuracies of >70% (LR 71.31% (2.37 SD) - SVM 79.40% (2.50 SD)) and demonstrate a previously unobserved performance disparity. Across all classifiers females suffer from a higher false negative rate (FNR). Presently, RF and LR classifiers are reported as the most effective models, yet in our experiments they demonstrate the greatest FNR disparity (RF; -21.02%; LR; -24.07%).

DISCUSSION

We demonstrate a sex disparity that exists in published ILPD classifiers. In practice, the higher FNR for females would manifest as increased rates of missed diagnosis for female patients and a consequent lack of appropriate care. Our study demonstrates that evaluating biases in the initial stages of machine learning can provide insights into inequalities in current clinical practice, reveal pathophysiological differences between the male and females, and can mitigate the digitisation of inequalities into algorithmic systems.

CONCLUSION

Our findings are important to medical data scientists, clinicians and policy-makers involved in the implementation medical artificial intelligence systems. An awareness of the potential biases of these systems is essential in preventing the digital exacerbation of healthcare inequalities.

摘要

目的

印度肝病患者数据集(ILPD)被广泛用于创建预测肝病的算法。鉴于现有的研究描述了肝病诊断和管理方面的人口统计学差异,这些算法需要仔细检查是否存在潜在偏差。我们通过研究 ILPD 模型中的性别偏差来解决这个被忽视的问题。

方法

在对 ILPD 论文进行文献回顾后,我们重新创建了现有研究中报告的模型,并对其进行了偏差检测。我们定义了四个实验,在性别不平衡/平衡数据上进行训练,并带有/不带有特征选择。我们构建了随机森林(RF)、支持向量机(SVM)、高斯朴素贝叶斯和逻辑回归(LR)分类器,进行了 100 次实验,报告平均结果及其标准差。

结果

我们复制了发表的模型,其准确率超过 70%(LR 为 71.31%(2.37 标准差)-SVM 为 79.40%(2.50 标准差)),并展示了一个以前未观察到的性能差异。在所有分类器中,女性的假阴性率(FNR)更高。目前,RF 和 LR 分类器被报告为最有效的模型,但在我们的实验中,它们表现出最大的 FNR 差异(RF:-21.02%;LR:-24.07%)。

讨论

我们展示了发表的 ILPD 分类器中存在的性别差异。在实践中,女性的 FNR 较高将表现为女性患者的漏诊率增加,以及相应的缺乏适当护理。我们的研究表明,在机器学习的初始阶段评估偏差可以深入了解当前临床实践中的不平等现象,揭示男性和女性之间的生理差异,并可以减轻不平等现象在算法系统中的数字化。

结论

我们的研究结果对参与实施医疗人工智能系统的医学数据科学家、临床医生和政策制定者非常重要。了解这些系统的潜在偏差对于防止医疗保健不平等现象的数字化加剧至关重要。

相似文献

引用本文的文献

5
The Limitations of Artificial Intelligence in Head and Neck Oncology.人工智能在头颈肿瘤学中的局限性
Adv Ther. 2025 Jun;42(6):2559-2568. doi: 10.1007/s12325-025-03198-4. Epub 2025 Apr 29.

本文引用的文献

1
Health inequities and the inappropriate use of race in nephrology.健康不公平和肾脏病学中种族的不当使用。
Nat Rev Nephrol. 2022 Feb;18(2):84-94. doi: 10.1038/s41581-021-00501-8. Epub 2021 Nov 8.
8
Machine learning in medicine: a practical introduction.医学中的机器学习:实用入门
BMC Med Res Methodol. 2019 Mar 19;19(1):64. doi: 10.1186/s12874-019-0681-4.

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验