• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于手机的心理健康评估算法中的公平性:探索性研究。

Fairness in Mobile Phone-Based Mental Health Assessment Algorithms: Exploratory Study.

作者信息

Park Jinkyung, Arunachalam Ramanathan, Silenzio Vincent, Singh Vivek K

机构信息

School of Communication & Information, Rutgers University, New Brunswick, NJ, United States.

Department of Computer Science, Rutgers University, New Brunswick, NJ, United States.

出版信息

JMIR Form Res. 2022 Jun 14;6(6):e34366. doi: 10.2196/34366.

DOI:10.2196/34366
PMID:35699997
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9240929/
Abstract

BACKGROUND

Approximately 1 in 5 American adults experience mental illness every year. Thus, mobile phone-based mental health prediction apps that use phone data and artificial intelligence techniques for mental health assessment have become increasingly important and are being rapidly developed. At the same time, multiple artificial intelligence-related technologies (eg, face recognition and search results) have recently been reported to be biased regarding age, gender, and race. This study moves this discussion to a new domain: phone-based mental health assessment algorithms. It is important to ensure that such algorithms do not contribute to gender disparities through biased predictions across gender groups.

OBJECTIVE

This research aimed to analyze the susceptibility of multiple commonly used machine learning approaches for gender bias in mobile mental health assessment and explore the use of an algorithmic disparate impact remover (DIR) approach to reduce bias levels while maintaining high accuracy.

METHODS

First, we performed preprocessing and model training using the data set (N=55) obtained from a previous study. Accuracy levels and differences in accuracy across genders were computed using 5 different machine learning models. We selected the random forest model, which yielded the highest accuracy, for a more detailed audit and computed multiple metrics that are commonly used for fairness in the machine learning literature. Finally, we applied the DIR approach to reduce bias in the mental health assessment algorithm.

RESULTS

The highest observed accuracy for the mental health assessment was 78.57%. Although this accuracy level raises optimism, the audit based on gender revealed that the performance of the algorithm was statistically significantly different between the male and female groups (eg, difference in accuracy across genders was 15.85%; P<.001). Similar trends were obtained for other fairness metrics. This disparity in performance was found to reduce significantly after the application of the DIR approach by adapting the data used for modeling (eg, the difference in accuracy across genders was 1.66%, and the reduction is statistically significant with P<.001).

CONCLUSIONS

This study grounds the need for algorithmic auditing in phone-based mental health assessment algorithms and the use of gender as a protected attribute to study fairness in such settings. Such audits and remedial steps are the building blocks for the widespread adoption of fair and accurate mental health assessment algorithms in the future.

摘要

背景

每年约五分之一的美国成年人患有精神疾病。因此,利用手机数据和人工智能技术进行心理健康评估的基于手机的心理健康预测应用程序变得越来越重要,并且正在迅速发展。与此同时,最近有报道称,多种与人工智能相关的技术(如人脸识别和搜索结果)在年龄、性别和种族方面存在偏差。本研究将这一讨论引入一个新领域:基于手机的心理健康评估算法。确保此类算法不会因跨性别群体的偏差预测而加剧性别差异非常重要。

目的

本研究旨在分析多种常用机器学习方法在移动心理健康评估中对性别偏差的敏感性,并探索使用算法差异影响消除器(DIR)方法在保持高精度的同时降低偏差水平。

方法

首先,我们使用从先前研究中获得的数据集(N = 55)进行预处理和模型训练。使用5种不同的机器学习模型计算准确率水平以及不同性别之间的准确率差异。我们选择了准确率最高的随机森林模型进行更详细的审查,并计算了机器学习文献中常用的多个公平性指标。最后,我们应用DIR方法来减少心理健康评估算法中的偏差。

结果

心理健康评估观察到的最高准确率为78.57%。尽管这一准确率水平令人乐观,但基于性别的审查表明,算法在男性和女性群体之间的表现存在统计学上的显著差异(例如,不同性别之间的准确率差异为15.85%;P <.001)。其他公平性指标也呈现类似趋势。通过调整用于建模的数据,发现应用DIR方法后,这种性能差异显著降低(例如,不同性别之间的准确率差异为1.66%,且降低具有统计学意义,P <.001)。

结论

本研究强调了在基于手机的心理健康评估算法中进行算法审查以及将性别作为受保护属性来研究此类环境中公平性的必要性。此类审查和补救措施是未来广泛采用公平准确的心理健康评估算法的基石。

相似文献

1
Fairness in Mobile Phone-Based Mental Health Assessment Algorithms: Exploratory Study.基于手机的心理健康评估算法中的公平性:探索性研究。
JMIR Form Res. 2022 Jun 14;6(6):e34366. doi: 10.2196/34366.
2
Evaluating and mitigating bias in machine learning models for cardiovascular disease prediction.评估和减轻心血管疾病预测机器学习模型中的偏差。
J Biomed Inform. 2023 Feb;138:104294. doi: 10.1016/j.jbi.2023.104294. Epub 2023 Jan 24.
3
Comparison of Methods to Reduce Bias From Clinical Prediction Models of Postpartum Depression.比较降低产后抑郁临床预测模型偏倚的方法。
JAMA Netw Open. 2021 Apr 1;4(4):e213909. doi: 10.1001/jamanetworkopen.2021.3909.
4
Evaluation and Mitigation of Racial Bias in Clinical Machine Learning Models: Scoping Review.临床机器学习模型中种族偏见的评估与缓解:范围综述
JMIR Med Inform. 2022 May 31;10(5):e36388. doi: 10.2196/36388.
5
A Racially Unbiased, Machine Learning Approach to Prediction of Mortality: Algorithm Development Study.一种基于机器学习的种族公平死亡率预测方法:算法开发研究。
JMIR Public Health Surveill. 2020 Oct 22;6(4):e22400. doi: 10.2196/22400.
6
A roadmap to artificial intelligence (AI): Methods for designing and building AI ready data to promote fairness.人工智能(AI)路线图:设计和构建 AI 就绪数据的方法,以促进公平性。
J Biomed Inform. 2024 Jun;154:104654. doi: 10.1016/j.jbi.2024.104654. Epub 2024 May 11.
7
Leveraging Mobile Phone Sensors, Machine Learning, and Explainable Artificial Intelligence to Predict Imminent Same-Day Binge-drinking Events to Support Just-in-time Adaptive Interventions: Algorithm Development and Validation Study.利用手机传感器、机器学习和可解释人工智能预测即将发生的当日暴饮事件以支持即时自适应干预:算法开发与验证研究
JMIR Form Res. 2023 May 4;7:e39862. doi: 10.2196/39862.
8
Testing Suicide Risk Prediction Algorithms Using Phone Measurements With Patients in Acute Mental Health Settings: Feasibility Study.使用急性心理健康环境中的患者的电话测量值测试自杀风险预测算法:可行性研究。
JMIR Mhealth Uhealth. 2020 Jun 26;8(6):e15901. doi: 10.2196/15901.
9
A joint fairness model with applications to risk predictions for underrepresented populations.具有代表性不足人群风险预测应用的联合公平模型。
Biometrics. 2023 Jun;79(2):826-840. doi: 10.1111/biom.13632. Epub 2022 Mar 27.
10
Sex-Based Performance Disparities in Machine Learning Algorithms for Cardiac Disease Prediction: Exploratory Study.基于性别的机器学习算法在心脏疾病预测中的表现差异:探索性研究。
J Med Internet Res. 2024 Aug 26;26:e46936. doi: 10.2196/46936.

引用本文的文献

1
Perspectives of Hispanic and Latinx Community Members on AI-Enabled mHealth Tools: Qualitative Focus Group Study.西班牙裔和拉丁裔社区成员对人工智能驱动的移动健康工具的看法:定性焦点小组研究。
J Med Internet Res. 2025 Feb 6;27:e59817. doi: 10.2196/59817.
2
Bias Mitigation in Primary Health Care Artificial Intelligence Models: Scoping Review.初级卫生保健人工智能模型中的偏差缓解:范围综述
J Med Internet Res. 2025 Jan 7;27:e60269. doi: 10.2196/60269.
3
Evaluating and mitigating unfairness in multimodal remote mental health assessments.评估和减轻多模式远程心理健康评估中的不公平现象。
PLOS Digit Health. 2024 Jul 24;3(7):e0000413. doi: 10.1371/journal.pdig.0000413. eCollection 2024 Jul.
4
Machine Learning for Multimodal Mental Health Detection: A Systematic Review of Passive Sensing Approaches.机器学习在多模态心理健康检测中的应用:被动感知方法的系统评价。
Sensors (Basel). 2024 Jan 6;24(2):348. doi: 10.3390/s24020348.
5
From quality improvement to equality improvement projects: A scoping review and framework.从质量改进到平等改进项目:一项范围审查与框架
iScience. 2023 Sep 15;26(10):107924. doi: 10.1016/j.isci.2023.107924. eCollection 2023 Oct 20.

本文引用的文献

1
A Racially Unbiased, Machine Learning Approach to Prediction of Mortality: Algorithm Development Study.一种基于机器学习的种族公平死亡率预测方法:算法开发研究。
JMIR Public Health Surveill. 2020 Oct 22;6(4):e22400. doi: 10.2196/22400.
2
Digital phenotyping for mental health of college students: a clinical review.大学生心理健康的数字化表型:临床综述。
Evid Based Ment Health. 2020 Nov;23(4):161-166. doi: 10.1136/ebmental-2020-300180. Epub 2020 Sep 30.
3
Dissecting racial bias in an algorithm used to manage the health of populations.剖析用于管理人群健康的算法中的种族偏见。
Science. 2019 Oct 25;366(6464):447-453. doi: 10.1126/science.aax2342.
4
Potential Biases in Machine Learning Algorithms Using Electronic Health Record Data.利用电子健康记录数据的机器学习算法中的潜在偏差。
JAMA Intern Med. 2018 Nov 1;178(11):1544-1547. doi: 10.1001/jamainternmed.2018.3763.
5
Fair Prediction with Disparate Impact: A Study of Bias in Recidivism Prediction Instruments.公平预测与差异影响:累犯预测工具中的偏见研究。
Big Data. 2017 Jun;5(2):153-163. doi: 10.1089/big.2016.0047.
6
Factors influencing the underutilization of mental health services among Asian American women with a history of depression and suicide.影响有抑郁和自杀史的亚裔美国女性心理健康服务利用不足的因素。
BMC Health Serv Res. 2015 Dec 8;15:542. doi: 10.1186/s12913-015-1191-7.
7
Mobile Phone Sensor Correlates of Depressive Symptom Severity in Daily-Life Behavior: An Exploratory Study.日常生活行为中手机传感器与抑郁症状严重程度的相关性:一项探索性研究。
J Med Internet Res. 2015 Jul 15;17(7):e175. doi: 10.2196/jmir.4273.
8
The MOS 36-item short-form health survey (SF-36). I. Conceptual framework and item selection.医学结局研究36项简明健康调查(SF-36)。I.概念框架与条目选择。
Med Care. 1992 Jun;30(6):473-83.