• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

ChatGPT 与医学专业人士在社交媒体上对检验医学问题的回答分析。

ChatGPT vs Medical Professional: Analyzing Responses to Laboratory Medicine Questions on Social Media.

机构信息

Department of Pathology, University of Michigan, Ann Arbor, MI, United States.

Department of Laboratory Medicine and Pathology, University of Washington, Seattle, WA, United States.

出版信息

Clin Chem. 2024 Sep 3;70(9):1122-1139. doi: 10.1093/clinchem/hvae093.

DOI:10.1093/clinchem/hvae093
PMID:39013110
Abstract

BACKGROUND

The integration of ChatGPT, a large language model (LLM) developed by OpenAI, into healthcare has sparked significant interest due to its potential to enhance patient care and medical education. With the increasing trend of patients accessing laboratory results online, there is a pressing need to evaluate the effectiveness of ChatGPT in providing accurate laboratory medicine information. Our study evaluates ChatGPT's effectiveness in addressing patient questions in this area, comparing its performance with that of medical professionals on social media.

METHODS

This study sourced patient questions and medical professional responses from Reddit and Quora, comparing them with responses generated by ChatGPT versions 3.5 and 4.0. Experienced laboratory medicine professionals evaluated the responses for quality and preference. Evaluation results were further analyzed using R software.

RESULTS

The study analyzed 49 questions, with evaluators reviewing responses from both medical professionals and ChatGPT. ChatGPT's responses were preferred by 75.9% of evaluators and generally received higher ratings for quality. They were noted for their comprehensive and accurate information, whereas responses from medical professionals were valued for their conciseness. The interrater agreement was fair, indicating some subjectivity but a consistent preference for ChatGPT's detailed responses.

CONCLUSIONS

ChatGPT demonstrates potential as an effective tool for addressing queries in laboratory medicine, often surpassing medical professionals in response quality. These results support the need for further research to confirm ChatGPT's utility and explore its integration into healthcare settings.

摘要

背景

OpenAI 开发的大型语言模型(LLM)ChatGPT 在医疗保健领域的整合引起了极大的兴趣,因为它有可能改善患者护理和医学教育。随着越来越多的患者在线访问实验室结果,迫切需要评估 ChatGPT 在提供准确的实验室医学信息方面的有效性。我们的研究评估了 ChatGPT 在解决这一领域患者问题方面的有效性,将其性能与社交媒体上的医疗专业人员进行了比较。

方法

本研究从 Reddit 和 Quora 上获取了患者问题和医学专业人员的回复,并将其与 ChatGPT 版本 3.5 和 4.0 生成的回复进行了比较。经验丰富的实验室医学专业人员评估了回复的质量和偏好。使用 R 软件进一步分析了评估结果。

结果

本研究分析了 49 个问题,评估人员审查了来自医学专业人员和 ChatGPT 的回复。75.9%的评估人员更喜欢 ChatGPT 的回复,并且总体上对其质量给予了更高的评价。ChatGPT 的回复以其全面和准确的信息为特点,而医学专业人员的回复则因其简洁而受到重视。评分者间的一致性为中等,表明存在一定的主观性,但对 ChatGPT 详细回复的偏好是一致的。

结论

ChatGPT 显示出在解决实验室医学查询方面具有潜力,其回复质量通常超过医学专业人员。这些结果支持进一步研究的必要性,以确认 ChatGPT 的实用性,并探索其在医疗保健环境中的整合。

相似文献

1
ChatGPT vs Medical Professional: Analyzing Responses to Laboratory Medicine Questions on Social Media.ChatGPT 与医学专业人士在社交媒体上对检验医学问题的回答分析。
Clin Chem. 2024 Sep 3;70(9):1122-1139. doi: 10.1093/clinchem/hvae093.
2
ChatGPT's performance in German OB/GYN exams - paving the way for AI-enhanced medical education and clinical practice.ChatGPT在德国妇产科考试中的表现——为人工智能强化医学教育和临床实践铺平道路。
Front Med (Lausanne). 2023 Dec 13;10:1296615. doi: 10.3389/fmed.2023.1296615. eCollection 2023.
3
Performance of ChatGPT on the Chinese Postgraduate Examination for Clinical Medicine: Survey Study.ChatGPT 在临床医学研究生入学考试中的表现:调查研究。
JMIR Med Educ. 2024 Feb 9;10:e48514. doi: 10.2196/48514.
4
Assessing the Quality of ChatGPT Responses to Dementia Caregivers' Questions: Qualitative Analysis.评估 ChatGPT 对痴呆症护理人员问题回答的质量:定性分析。
JMIR Aging. 2024 May 6;7:e53019. doi: 10.2196/53019.
5
Evaluating ChatGPT's effectiveness and tendencies in Japanese internal medicine.评估 ChatGPT 在日本内科学中的有效性和倾向。
J Eval Clin Pract. 2024 Sep;30(6):1017-1023. doi: 10.1111/jep.14011. Epub 2024 May 19.
6
Evaluation of ChatGPT's responses to information needs and information seeking of dementia patients.评估 ChatGPT 对痴呆症患者信息需求和信息检索的响应。
Sci Rep. 2024 May 4;14(1):10273. doi: 10.1038/s41598-024-61068-5.
7
A Multidisciplinary Assessment of ChatGPT's Knowledge of Amyloidosis: Observational Study.对ChatGPT关于淀粉样变性知识的多学科评估:观察性研究。
JMIR Cardio. 2024 Apr 19;8:e53421. doi: 10.2196/53421.
8
ChatGPT Versus Consultants: Blinded Evaluation on Answering Otorhinolaryngology Case-Based Questions.ChatGPT与医学顾问的对比:对耳鼻喉科基于病例问题回答的盲法评估
JMIR Med Educ. 2023 Dec 5;9:e49183. doi: 10.2196/49183.
9
Quality and Dependability of ChatGPT and DingXiangYuan Forums for Remote Orthopedic Consultations: Comparative Analysis.ChatGPT 和丁香园论坛在远程骨科咨询中的质量和可靠性:比较分析。
J Med Internet Res. 2024 Mar 14;26:e50882. doi: 10.2196/50882.
10
Assessing ChatGPT's capacity for clinical decision support in pediatrics: A comparative study with pediatricians using KIDMAP of Rasch analysis.评估 ChatGPT 在儿科临床决策支持方面的能力:使用 KIDMAP 的 Rasch 分析与儿科医生进行的比较研究。
Medicine (Baltimore). 2023 Jun 23;102(25):e34068. doi: 10.1097/MD.0000000000034068.

引用本文的文献

1
Evaluating Large Language Models for Preoperative Patient Education in Superior Capsular Reconstruction: Comparative Study of Claude, GPT, and Gemini.评估大语言模型在肩胛下肌上囊重建术前患者教育中的应用:Claude、GPT和Gemini的比较研究
JMIR Perioper Med. 2025 Jun 12;8:e70047. doi: 10.2196/70047.
2
LabQAR: A Manually Curated Dataset for Question Answering on Laboratory Test Reference Ranges and Interpretation.LabQAR:一个人工整理的关于实验室检查参考范围及解读问答的数据集。
medRxiv. 2025 Jun 3:2025.06.03.25328882. doi: 10.1101/2025.06.03.25328882.
3
Performance evaluation of large language models with chain-of-thought reasoning ability in clinical laboratory case interpretation.
具有思维链推理能力的大语言模型在临床检验病例解读中的性能评估
Clin Chem Lab Med. 2025 Mar 3. doi: 10.1515/cclm-2025-0055.