• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

相似文献

1
Evaluating for Evidence of Sociodemographic Bias in Conversational AI for Mental Health Support.评估用于心理健康支持的对话式人工智能中社会人口统计学偏见的证据。
Cyberpsychol Behav Soc Netw. 2025 Jan;28(1):44-51. doi: 10.1089/cyber.2024.0199. Epub 2024 Oct 24.
2
Interventions for central serous chorioretinopathy: a network meta-analysis.中心性浆液性脉络膜视网膜病变的干预措施:一项网状Meta分析
Cochrane Database Syst Rev. 2025 Jun 16;6(6):CD011841. doi: 10.1002/14651858.CD011841.pub3.
3
Home treatment for mental health problems: a systematic review.心理健康问题的居家治疗:一项系统综述
Health Technol Assess. 2001;5(15):1-139. doi: 10.3310/hta5150.
4
Signs and symptoms to determine if a patient presenting in primary care or hospital outpatient settings has COVID-19.在基层医疗机构或医院门诊环境中,如果患者出现以下症状和体征,可判断其是否患有 COVID-19。
Cochrane Database Syst Rev. 2022 May 20;5(5):CD013665. doi: 10.1002/14651858.CD013665.pub3.
5
A New Measure of Quantified Social Health Is Associated With Levels of Discomfort, Capability, and Mental and General Health Among Patients Seeking Musculoskeletal Specialty Care.一种新的量化社会健康指标与寻求肌肉骨骼专科护理的患者的不适程度、能力以及心理和总体健康水平相关。
Clin Orthop Relat Res. 2025 Apr 1;483(4):647-663. doi: 10.1097/CORR.0000000000003394. Epub 2025 Feb 5.
6
Falls prevention interventions for community-dwelling older adults: systematic review and meta-analysis of benefits, harms, and patient values and preferences.社区居住的老年人跌倒预防干预措施:系统评价和荟萃分析的益处、危害以及患者的价值观和偏好。
Syst Rev. 2024 Nov 26;13(1):289. doi: 10.1186/s13643-024-02681-3.
7
Communication skills training for mental health professionals working with people with severe mental illness.针对为重度精神疾病患者提供服务的心理健康专业人员的沟通技能培训。
Cochrane Database Syst Rev. 2017 Jun 13;6(6):CD010006. doi: 10.1002/14651858.CD010006.pub2.
8
Eliciting adverse effects data from participants in clinical trials.从临床试验参与者中获取不良反应数据。
Cochrane Database Syst Rev. 2018 Jan 16;1(1):MR000039. doi: 10.1002/14651858.MR000039.pub2.
9
Interventions for improving adherence to amblyopia treatments in children.改善儿童弱视治疗依从性的干预措施。
Cochrane Database Syst Rev. 2025 Jul 2;7(7):CD015820. doi: 10.1002/14651858.CD015820.pub2.
10
Control interventions in randomised trials among people with mental health disorders.精神障碍患者随机试验中的对照干预措施。
Cochrane Database Syst Rev. 2022 Apr 4;4(4):MR000050. doi: 10.1002/14651858.MR000050.pub2.

本文引用的文献

1
Performance of ChatGPT as an AI-assisted decision support tool in medicine: a proof-of-concept study for interpreting symptoms and management of common cardiac conditions (AMSTELHEART-2).ChatGPT 在医学中作为 AI 辅助决策支持工具的性能:解释常见心脏疾病症状和管理的概念验证研究 (AMSTELHEART-2)。
Acta Cardiol. 2024 May;79(3):358-366. doi: 10.1080/00015385.2024.2303528. Epub 2024 Feb 13.
2
Feasibility of combining spatial computing and AI for mental health support in anxiety and depression.将空间计算与人工智能相结合用于焦虑和抑郁心理健康支持的可行性。
NPJ Digit Med. 2024 Jan 26;7(1):22. doi: 10.1038/s41746-024-01011-0.
3
The Accuracy and Potential Racial and Ethnic Biases of GPT-4 in the Diagnosis and Triage of Health Conditions: Evaluation Study.GPT-4在健康状况诊断和分诊中的准确性及潜在的种族和民族偏见:评估研究
JMIR Med Educ. 2023 Nov 2;9:e47532. doi: 10.2196/47532.
4
Large language models propagate race-based medicine.大语言模型传播基于种族的医学观念。
NPJ Digit Med. 2023 Oct 20;6(1):195. doi: 10.1038/s41746-023-00939-z.
5
Applications of ChatGPT and Large Language Models in Medicine and Health Care: Benefits and Pitfalls.ChatGPT和大语言模型在医学与医疗保健中的应用:益处与风险
Fed Pract. 2023 Jun;40(6):170-173. doi: 10.12788/fp.0386. Epub 2023 Jun 16.
6
Accuracy and Reliability of Chatbot Responses to Physician Questions.聊天机器人对医生提问回答的准确性和可靠性。
JAMA Netw Open. 2023 Oct 2;6(10):e2336483. doi: 10.1001/jamanetworkopen.2023.36483.
7
The Role of Large Language Models in Medical Education: Applications and Implications.大语言模型在医学教育中的作用:应用与启示
JMIR Med Educ. 2023 Aug 14;9:e50945. doi: 10.2196/50945.
8
Sentiment analysis: A survey on design framework, applications and future scopes.情感分析:关于设计框架、应用及未来发展范围的综述
Artif Intell Rev. 2023 Mar 20:1-56. doi: 10.1007/s10462-023-10442-2.
9
Artificial intelligence and anaesthesia examinations: exploring ChatGPT as a prelude to the future.人工智能与麻醉学考试:探索ChatGPT作为未来的前奏
Br J Anaesth. 2023 Aug;131(2):e36-e37. doi: 10.1016/j.bja.2023.04.033. Epub 2023 May 26.
10
Appropriateness of Breast Cancer Prevention and Screening Recommendations Provided by ChatGPT.ChatGPT提供的乳腺癌预防和筛查建议的适宜性。
Radiology. 2023 May;307(4):e230424. doi: 10.1148/radiol.230424. Epub 2023 Apr 4.

评估用于心理健康支持的对话式人工智能中社会人口统计学偏见的证据。

Evaluating for Evidence of Sociodemographic Bias in Conversational AI for Mental Health Support.

作者信息

Yeo Yee Hui, Peng Yuxin, Mehra Muskaan, Samaan Jamil, Hakimian Joshua, Clark Allistair, Suchak Karisma, Krut Zoe, Andersson Taiga, Persky Susan, Liran Omer, Spiegel Brennan

机构信息

Karsh Division of Gastroenterology and Hepatology, Cedars-Sinai Medical Center, Los Angeles, California, USA.

School of Mathematics and Statistics, Xi'an Jiaotong University, Xi'an, China.

出版信息

Cyberpsychol Behav Soc Netw. 2025 Jan;28(1):44-51. doi: 10.1089/cyber.2024.0199. Epub 2024 Oct 24.

DOI:10.1089/cyber.2024.0199
PMID:39446671
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11807910/
Abstract

The integration of large language models (LLMs) into healthcare highlights the need to ensure their efficacy while mitigating potential harms, such as the perpetuation of biases. Current evidence on the existence of bias within LLMs remains inconclusive. In this study, we present an approach to investigate the presence of bias within an LLM designed for mental health support. We simulated physician-patient conversations by using a communication loop between an LLM-based conversational agent and digital standardized patients (DSPs) that engaged the agent in dialogue while remaining agnostic to sociodemographic characteristics. In contrast, the conversational agent was made aware of each DSP's characteristics, including age, sex, race/ethnicity, and annual income. The agent's responses were analyzed to discern potential systematic biases using the Linguistic Inquiry and Word Count tool. Multivariate regression analysis, trend analysis, and group-based trajectory models were used to quantify potential biases. Among 449 conversations, there was no evidence of bias in both descriptive assessments and multivariable linear regression analyses. Moreover, when evaluating changes in mean tone scores throughout a dialogue, the conversational agent exhibited a capacity to show understanding of the DSPs' chief complaints and to elevate the tone scores of the DSPs throughout conversations. This finding did not vary by any sociodemographic characteristics of the DSP. Using an objective methodology, our study did not uncover significant evidence of bias within an LLM-enabled mental health conversational agent. These findings offer a complementary approach to examining bias in LLM-based conversational agents for mental health support.

摘要

将大语言模型(LLMs)整合到医疗保健领域凸显了在确保其有效性的同时减轻潜在危害的必要性,比如避免偏见的持续存在。目前关于大语言模型中偏见存在与否的证据尚无定论。在本研究中,我们提出了一种方法来调查一个用于心理健康支持的大语言模型中是否存在偏见。我们通过基于大语言模型的对话代理与数字标准化患者(DSPs)之间的通信循环来模拟医患对话,数字标准化患者在与代理对话时对社会人口统计学特征保持未知。相比之下,对话代理知晓每个DSP的特征,包括年龄、性别、种族/民族和年收入。使用语言查询与字数统计工具分析代理的回复,以辨别潜在的系统偏见。采用多元回归分析、趋势分析和基于组的轨迹模型来量化潜在偏见。在449次对话中,描述性评估和多变量线性回归分析均未发现偏见证据。此外,在评估整个对话过程中平均语气得分的变化时,对话代理表现出能够理解DSP的主要诉求,并在整个对话过程中提高DSP的语气得分。这一发现不因DSP的任何社会人口统计学特征而有所不同。通过一种客观的方法,我们的研究未发现基于大语言模型的心理健康对话代理存在显著偏见证据。这些发现为检查基于大语言模型的心理健康支持对话代理中的偏见提供了一种补充方法。