• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

剖析数字健康领域基于语音的机器学习模型中的人口统计学偏差

Deconstructing demographic bias in speech-based machine learning models for digital health.

作者信息

Yang Michael, El-Attar Abd-Allah, Chaspari Theodora

机构信息

Computer Science & Engineering, Texas A&M University, College Station, TX, United States.

Computer Science & Engineering, Texas A&M University Qatar, Al Rayyan, Qatar.

出版信息

Front Digit Health. 2024 Jul 25;6:1351637. doi: 10.3389/fdgth.2024.1351637. eCollection 2024.

DOI:10.3389/fdgth.2024.1351637
PMID:39119589
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11306200/
Abstract

INTRODUCTION

Machine learning (ML) algorithms have been heralded as promising solutions to the realization of assistive systems in digital healthcare, due to their ability to detect fine-grain patterns that are not easily perceived by humans. Yet, ML algorithms have also been critiqued for treating individuals differently based on their demography, thus propagating existing disparities. This paper explores gender and race bias in speech-based ML algorithms that detect behavioral and mental health outcomes.

METHODS

This paper examines potential sources of bias in the data used to train the ML, encompassing acoustic features extracted from speech signals and associated labels, as well as in the ML decisions. The paper further examines approaches to reduce existing bias via using the features that are the least informative of one's demographic information as the ML input, and transforming the feature space in an adversarial manner to diminish the evidence of the demographic information while retaining information about the focal behavioral and mental health state.

RESULTS

Results are presented in two domains, the first pertaining to gender and race bias when estimating levels of anxiety, and the second pertaining to gender bias in detecting depression. Findings indicate the presence of statistically significant differences in both acoustic features and labels among demographic groups, as well as differential ML performance among groups. The statistically significant differences present in the label space are partially preserved in the ML decisions. Although variations in ML performance across demographic groups were noted, results are mixed regarding the models' ability to accurately estimate healthcare outcomes for the sensitive groups.

DISCUSSION

These findings underscore the necessity for careful and thoughtful design in developing ML models that are capable of maintaining crucial aspects of the data and perform effectively across all populations in digital healthcare applications.

摘要

引言

机器学习(ML)算法被誉为实现数字医疗辅助系统的有前景的解决方案,因为它们能够检测人类不易察觉的细粒度模式。然而,ML算法也因根据人口统计学特征区别对待个体而受到批评,从而加剧了现有的差异。本文探讨了用于检测行为和心理健康结果的基于语音的ML算法中的性别和种族偏见。

方法

本文研究了用于训练ML的数据中潜在的偏见来源,包括从语音信号中提取的声学特征和相关标签,以及ML决策中的偏见来源。本文还研究了通过使用对个人人口统计信息最缺乏信息量的特征作为ML输入,并以对抗方式变换特征空间来减少现有偏见的方法,以减少人口统计信息的证据,同时保留有关重点行为和心理健康状态的信息。

结果

结果在两个领域呈现,第一个领域涉及估计焦虑水平时的性别和种族偏见,第二个领域涉及检测抑郁症时的性别偏见。研究结果表明,不同人口群体在声学特征和标签方面存在统计学上的显著差异,并且群体之间的ML性能也存在差异。标签空间中存在的统计学显著差异在ML决策中部分保留。尽管注意到不同人口群体的ML性能存在差异,但关于模型准确估计敏感群体医疗结果的能力,结果好坏参半。

讨论

这些发现强调了在开发ML模型时进行谨慎和深思熟虑设计的必要性,这些模型能够保留数据的关键方面,并在数字医疗应用中对所有人群有效发挥作用。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2e3/11306200/54533c434a25/fdgth-06-1351637-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2e3/11306200/54533c434a25/fdgth-06-1351637-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e2e3/11306200/54533c434a25/fdgth-06-1351637-g001.jpg

相似文献

1
Deconstructing demographic bias in speech-based machine learning models for digital health.剖析数字健康领域基于语音的机器学习模型中的人口统计学偏差
Front Digit Health. 2024 Jul 25;6:1351637. doi: 10.3389/fdgth.2024.1351637. eCollection 2024.
2
Evaluating and mitigating bias in machine learning models for cardiovascular disease prediction.评估和减轻心血管疾病预测机器学习模型中的偏差。
J Biomed Inform. 2023 Feb;138:104294. doi: 10.1016/j.jbi.2023.104294. Epub 2023 Jan 24.
3
Fairness in Mobile Phone-Based Mental Health Assessment Algorithms: Exploratory Study.基于手机的心理健康评估算法中的公平性:探索性研究。
JMIR Form Res. 2022 Jun 14;6(6):e34366. doi: 10.2196/34366.
4
Folic acid supplementation and malaria susceptibility and severity among people taking antifolate antimalarial drugs in endemic areas.在流行地区,服用抗叶酸抗疟药物的人群中,叶酸补充剂与疟疾易感性和严重程度的关系。
Cochrane Database Syst Rev. 2022 Feb 1;2(2022):CD014217. doi: 10.1002/14651858.CD014217.
5
Using Wearable Devices and Speech Data for Personalized Machine Learning in Early Detection of Mental Disorders: Protocol for a Participatory Research Study.利用可穿戴设备和语音数据进行精神障碍早期检测的个性化机器学习:一项参与性研究方案
JMIR Res Protoc. 2023 Nov 13;12:e48210. doi: 10.2196/48210.
6
7
Comparison of Methods to Reduce Bias From Clinical Prediction Models of Postpartum Depression.比较降低产后抑郁临床预测模型偏倚的方法。
JAMA Netw Open. 2021 Apr 1;4(4):e213909. doi: 10.1001/jamanetworkopen.2021.3909.
8
A roadmap to artificial intelligence (AI): Methods for designing and building AI ready data to promote fairness.人工智能(AI)路线图:设计和构建 AI 就绪数据的方法,以促进公平性。
J Biomed Inform. 2024 Jun;154:104654. doi: 10.1016/j.jbi.2024.104654. Epub 2024 May 11.
9
Digital Biomarkers for Depression Screening With Wearable Devices: Cross-sectional Study With Machine Learning Modeling.基于可穿戴设备的抑郁筛查数字生物标志物:机器学习建模的横断面研究。
JMIR Mhealth Uhealth. 2021 Oct 25;9(10):e24872. doi: 10.2196/24872.
10
Evolution and impact of bias in human and machine learning algorithm interaction.人类与机器学习算法交互中的偏差演变与影响。
PLoS One. 2020 Aug 13;15(8):e0235502. doi: 10.1371/journal.pone.0235502. eCollection 2020.

本文引用的文献

1
Manifestation of depression in speech overlaps with characteristics used to represent and recognize speaker identity.抑郁在言语中的表现与用于代表和识别说话者身份的特征重叠。
Sci Rep. 2023 Jul 10;13(1):11155. doi: 10.1038/s41598-023-35184-7.
2
Sources of bias in artificial intelligence that perpetuate healthcare disparities-A global review.导致医疗保健差距长期存在的人工智能偏差来源——一项全球综述。
PLOS Digit Health. 2022 Mar 31;1(3):e0000022. doi: 10.1371/journal.pdig.0000022. eCollection 2022 Mar.
3
A Call to Action on Assessing and Mitigating Bias in Artificial Intelligence Applications for Mental Health.
呼吁重视并减轻人工智能应用于精神健康领域中的偏见
Perspect Psychol Sci. 2023 Sep;18(5):1062-1096. doi: 10.1177/17456916221134490. Epub 2022 Dec 9.
4
Variation in global and intonational pitch settings among black and white speakers of Southern American English.美国南部英语中黑人和白人说话者在整体音高和语调音高设置上的差异。
J Acoust Soc Am. 2022 Nov;152(5):2617. doi: 10.1121/10.0014906.
5
Speech as an indicator for psychosocial stress: A network analytic approach.言语作为心理社会压力的指标:一种网络分析方法。
Behav Res Methods. 2022 Apr;54(2):910-921. doi: 10.3758/s13428-021-01670-x. Epub 2021 Aug 6.
6
Comparison of Methods to Reduce Bias From Clinical Prediction Models of Postpartum Depression.比较降低产后抑郁临床预测模型偏倚的方法。
JAMA Netw Open. 2021 Apr 1;4(4):e213909. doi: 10.1001/jamanetworkopen.2021.3909.
7
Accuracy comparison across face recognition algorithms: Where are we on measuring race bias?不同人脸识别算法的准确性比较:在衡量种族偏见方面我们处于什么位置?
IEEE Trans Biom Behav Identity Sci. 2021 Jan;3(1):101-111. doi: 10.1109/TBIOM.2020.3027269. Epub 2020 Sep 29.
8
Big data in digital healthcare: lessons learnt and recommendations for general practice.数字医疗中的大数据:全科医学的经验教训和建议。
Heredity (Edinb). 2020 Apr;124(4):525-534. doi: 10.1038/s41437-020-0303-2. Epub 2020 Mar 5.
9
Diagnosing bias in data-driven algorithms for healthcare.诊断医疗保健数据驱动算法中的偏差。
Nat Med. 2020 Jan;26(1):25-26. doi: 10.1038/s41591-019-0726-6.
10
Addressing Bias in Artificial Intelligence in Health Care.应对医疗保健领域人工智能中的偏见问题。
JAMA. 2019 Dec 24;322(24):2377-2378. doi: 10.1001/jama.2019.18058.