• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于元数据的医疗保健时间序列偏差分析(BAHT)决策支持系统

Bias Analysis in Healthcare Time Series (BAHT) Decision Support Systems from Meta Data.

作者信息

Dakshit Sagnik, Dakshit Sristi, Khargonkar Ninad, Prabhakaran Balakrishnan

机构信息

Computer Science, The University of Texas at Dallas, Dallas, USA.

出版信息

J Healthc Inform Res. 2023 Jun 19;7(2):225-253. doi: 10.1007/s41666-023-00133-6. eCollection 2023 Jun.

DOI:10.1007/s41666-023-00133-6
PMID:37377633
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10290973/
Abstract

One of the hindrances in the widespread acceptance of deep learning-based decision support systems in healthcare is bias. Bias in its many forms occurs in the datasets used to train and test deep learning models and is amplified when deployed in the real world, leading to challenges such as model drift. Recent advancements in the field of deep learning have led to the deployment of deployable automated healthcare diagnosis decision support systems at hospitals as well as tele-medicine through IoT devices. Research has been focused primarily on the development and improvement of these systems leaving a gap in the analysis of the fairness. The domain of FAccT ML (fairness, accountability, and transparency) accounts for the analysis of these deployable machine learning systems. In this work, we present a framework for bias analysis in healthcare time series (BAHT) signals such as electrocardiogram (ECG) and electroencephalogram (EEG). BAHT provides a graphical interpretive analysis of bias in the training, testing datasets in terms of protected variables, and analysis of bias amplification by the trained supervised learning model for time series healthcare decision support systems. We thoroughly investigate three prominent time series ECG and EEG healthcare datasets used for model training and research. We show the extensive presence of bias in the datasets leads to potentially biased or unfair machine-learning models. Our experiments also demonstrate the amplification of identified bias with an observed maximum of 66.66%. We investigate the effect of model drift due to unanalyzed bias in datasets and algorithms. Bias mitigation though prudent is a nascent area of research. We present experiments and analyze the most prevalently accepted bias mitigation strategies of under-sampling, oversampling, and the use of synthetic data for balancing the dataset through augmentation. It is important that healthcare models, datasets, and bias mitigation strategies should be properly analyzed for a fair unbiased delivery of service.

摘要

深度学习在医疗保健领域的决策支持系统广泛应用中面临的障碍之一是偏差。偏差以多种形式出现在用于训练和测试深度学习模型的数据集中,并且在实际应用中会被放大,从而导致模型漂移等问题。深度学习领域的最新进展促使可部署的自动化医疗诊断决策支持系统通过物联网设备在医院以及远程医疗中得到应用。研究主要集中在这些系统的开发和改进上,而在公平性分析方面存在空白。FAccT ML(公平性、问责制和透明度)领域负责对这些可部署的机器学习系统进行分析。在这项工作中,我们提出了一个针对医疗时间序列(BAHT)信号(如心电图(ECG)和脑电图(EEG))的偏差分析框架。BAHT从受保护变量的角度对训练和测试数据集中的偏差进行图形化解释分析,并对用于时间序列医疗决策支持系统的训练有监督学习模型的偏差放大情况进行分析。我们深入研究了用于模型训练和研究的三个著名的时间序列ECG和EEG医疗数据集。我们发现数据集中广泛存在的偏差会导致潜在的有偏差或不公平的机器学习模型。我们的实验还证明了所识别偏差的放大,观察到的最大放大率为66.66%。我们研究了由于数据集中未分析的偏差和算法导致的模型漂移的影响。尽管谨慎地减轻偏差是一个新兴的研究领域。我们展示了实验,并分析了最普遍接受的偏差减轻策略,即欠采样、过采样以及使用合成数据通过扩充来平衡数据集。重要的是,应该对医疗模型、数据集和偏差减轻策略进行适当分析,以实现公平无偏差的服务提供。

相似文献

1
Bias Analysis in Healthcare Time Series (BAHT) Decision Support Systems from Meta Data.基于元数据的医疗保健时间序列偏差分析(BAHT)决策支持系统
J Healthc Inform Res. 2023 Jun 19;7(2):225-253. doi: 10.1007/s41666-023-00133-6. eCollection 2023 Jun.
2
Enabling Fairness in Healthcare Through Machine Learning.通过机器学习实现医疗保健中的公平性。
Ethics Inf Technol. 2022;24(3):39. doi: 10.1007/s10676-022-09658-7. Epub 2022 Aug 31.
3
A scoping review of fair machine learning techniques when using real-world data.使用真实世界数据时公平机器学习技术的范围综述。
J Biomed Inform. 2024 Mar;151:104622. doi: 10.1016/j.jbi.2024.104622. Epub 2024 Mar 6.
4
Deep convolutional neural network and IoT technology for healthcare.用于医疗保健的深度卷积神经网络和物联网技术。
Digit Health. 2024 Jan 17;10:20552076231220123. doi: 10.1177/20552076231220123. eCollection 2024 Jan-Dec.
5
CEFEs: A CNN Explainable Framework for ECG Signals.CEFEs:用于心电图信号的 CNN 可解释框架。
Artif Intell Med. 2021 May;115:102059. doi: 10.1016/j.artmed.2021.102059. Epub 2021 Mar 26.
6
D-BIAS: A Causality-Based Human-in-the-Loop System for Tackling Algorithmic Bias.D-BIAS:一种基于因果关系的人在回路系统,用于解决算法偏差。
IEEE Trans Vis Comput Graph. 2023 Jan;29(1):473-482. doi: 10.1109/TVCG.2022.3209484. Epub 2022 Dec 16.
7
A novel approach for assessing fairness in deployed machine learning algorithms.一种评估已部署机器学习算法公平性的新方法。
Sci Rep. 2024 Aug 1;14(1):17753. doi: 10.1038/s41598-024-68651-w.
8
Assessment of differentially private synthetic data for utility and fairness in end-to-end machine learning pipelines for tabular data.用于表格数据的端到端机器学习管道中效用和公平性的差分隐私合成数据评估。
PLoS One. 2024 Feb 5;19(2):e0297271. doi: 10.1371/journal.pone.0297271. eCollection 2024.
9
The future of Cochrane Neonatal.考克兰新生儿协作网的未来。
Early Hum Dev. 2020 Nov;150:105191. doi: 10.1016/j.earlhumdev.2020.105191. Epub 2020 Sep 12.
10
BoostedEnML: Efficient Technique for Detecting Cyberattacks in IoT Systems Using Boosted Ensemble Machine Learning.增强集成机器学习的物联网系统网络攻击检测高效技术:BoostedEnML
Sensors (Basel). 2022 Sep 29;22(19):7409. doi: 10.3390/s22197409.

引用本文的文献

1
Uncovering ethical biases in publicly available fetal ultrasound datasets.揭示公开可用的胎儿超声数据集中的伦理偏见。
NPJ Digit Med. 2025 Jun 13;8(1):355. doi: 10.1038/s41746-025-01739-3.
2
Synthetic data generation methods in healthcare: A review on open-source tools and methods.医疗保健领域的合成数据生成方法:关于开源工具和方法的综述
Comput Struct Biotechnol J. 2024 Jul 9;23:2892-2910. doi: 10.1016/j.csbj.2024.07.005. eCollection 2024 Dec.

本文引用的文献

1
Does imbalance in chest X-ray datasets produce biased deep learning approaches for COVID-19 screening?胸部 X 射线数据集的不平衡是否会导致用于 COVID-19 筛查的深度学习方法产生偏差?
BMC Med Res Methodol. 2022 Apr 28;22(1):125. doi: 10.1186/s12874-022-01578-w.
2
Electronic Health Records as Biased Tools or Tools Against Bias: A Conceptual Model.电子健康记录:偏见工具还是反偏见工具?一个概念模型。
Milbank Q. 2022 Mar;100(1):134-150. doi: 10.1111/1468-0009.12545. Epub 2021 Nov 23.
3
Public Covid-19 X-ray datasets and their impact on model bias - A systematic review of a significant problem.公开的新冠 X 射线数据集及其对模型偏差的影响 - 一个重大问题的系统综述。
Med Image Anal. 2021 Dec;74:102225. doi: 10.1016/j.media.2021.102225. Epub 2021 Sep 28.
4
The Problem of Fairness in Synthetic Healthcare Data.合成医疗数据中的公平性问题。
Entropy (Basel). 2021 Sep 4;23(9):1165. doi: 10.3390/e23091165.
5
CEFEs: A CNN Explainable Framework for ECG Signals.CEFEs:用于心电图信号的 CNN 可解释框架。
Artif Intell Med. 2021 May;115:102059. doi: 10.1016/j.artmed.2021.102059. Epub 2021 Mar 26.
6
Inherent Bias in Artificial Intelligence-Based Decision Support Systems for Healthcare.人工智能在医疗保健决策支持系统中的固有偏差。
Medicina (Kaunas). 2020 Mar 20;56(3):141. doi: 10.3390/medicina56030141.
7
Benefits, Pitfalls, and Potential Bias in Health Care AI.医疗保健人工智能的益处、陷阱与潜在偏差
N C Med J. 2019 Jul-Aug;80(4):219-223. doi: 10.18043/ncm.80.4.219.
8
Addressing Bias in Electronic Health Record-Based Surveillance of Cardiovascular Disease Risk: Finding the Signal Through the Noise.解决基于电子健康记录的心血管疾病风险监测中的偏差:透过噪声找到信号。
Curr Epidemiol Rep. 2017 Dec;4(4):346-352. doi: 10.1007/s40471-017-0130-z. Epub 2017 Nov 2.
9
Potential Biases in Machine Learning Algorithms Using Electronic Health Record Data.利用电子健康记录数据的机器学习算法中的潜在偏差。
JAMA Intern Med. 2018 Nov 1;178(11):1544-1547. doi: 10.1001/jamainternmed.2018.3763.
10
Comparing humans and deep learning performance for grading AMD: A study in using universal deep features and transfer learning for automated AMD analysis.比较人类与深度学习在年龄相关性黄斑变性分级方面的表现:一项关于使用通用深度特征和迁移学习进行年龄相关性黄斑变性自动分析的研究。
Comput Biol Med. 2017 Mar 1;82:80-86. doi: 10.1016/j.compbiomed.2017.01.018. Epub 2017 Jan 27.