文献检索文档翻译深度研究
Suppr Zotero 插件Zotero 插件
邀请有礼套餐&价格历史记录

新学期,新优惠

限时优惠:9月1日-9月22日

30天高级会员仅需29元

1天体验卡首发特惠仅需5.99元

了解详情
不再提醒
插件&应用
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
高级版
套餐订阅购买积分包
AI 工具
文献检索文档翻译深度研究
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2025

评估ChatGPT在提供儿科疾病治疗建议方面的表现。

Evaluation of ChatGPT's performance in providing treatment recommendations for pediatric diseases.

作者信息

Wei Qiuhong, Wang Yanqin, Yao Zhengxiong, Cui Ying, Wei Bo, Li Tingyu, Xu Ximing

机构信息

Children Nutrition Research Center Children's Hospital of Chongqing Medical University National Clinical Research Center for Child Health and Disorders Ministry of Education Key Laboratory of Child Development and Disorders China International Science and Technology Cooperation Base of Child Development and Critical Disorders Chongqing Key Laboratory of Childhood Nutrition and Health Chongqing China.

College of Medical Informatics Medical Data Science Academy Chongqing Engineering Research Center for Clinical Big-Data and Drug Evaluation Chongqing Medical University Chongqing China.

出版信息

Pediatr Discov. 2023 Nov 20;1(3):e42. doi: 10.1002/pdi3.42. eCollection 2023 Dec.


DOI:10.1002/pdi3.42
PMID:40626087
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12118210/
Abstract

With the advance of artificial intelligence technology, large language models such as ChatGPT are drawing substantial interest in the healthcare field. A growing body of research has evaluated ChatGPT's performance in various medical departments, yet its potential in pediatrics remains under-studied. In this study, we presented ChatGPT with a total of 4160 clinical consultation questions in both English and Chinese, covering 104 pediatric conditions, and repeated each question independently 10 times to assess the accuracy of its responses in pediatric disease treatment recommendations. ChatGPT achieved an overall accuracy of 82.2% (95% CI: 81.0%-83.4%), with superior performance in addressing common diseases (84.4%, 95% CI: 83.2%-85.7%), offering general treatment advice (83.5%, 95% CI: 81.9%-85.1%), and responding in English (93.0%, 95% CI: 91.9%-94.1%). However, it was prone to errors in disease definitions, medications, and surgical treatment. In conclusion, while ChatGPT shows promise in pediatric treatment recommendations with notable accuracy, cautious optimism is warranted regarding the potential application of large language models in enhancing patient care.

摘要

随着人工智能技术的发展,ChatGPT等大型语言模型在医疗保健领域引起了广泛关注。越来越多的研究评估了ChatGPT在各个医学科室的表现,但其在儿科的潜力仍研究不足。在本研究中,我们向ChatGPT提出了总共4160个中英文临床咨询问题,涵盖104种儿科疾病,并对每个问题独立重复10次,以评估其在儿科疾病治疗建议中回答的准确性。ChatGPT的总体准确率为82.2%(95%置信区间:81.0%-83.4%),在处理常见疾病(84.4%,95%置信区间:83.2%-85.7%)、提供一般治疗建议(83.5%,95%置信区间:81.9%-85.1%)以及用英语回答(93.0%,95%置信区间:91.9%-94.1%)方面表现出色。然而,它在疾病定义、药物治疗和手术治疗方面容易出错。总之,虽然ChatGPT在儿科治疗建议方面显示出有前景的准确性,但对于大型语言模型在改善患者护理方面的潜在应用,仍需谨慎乐观。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e9b5/12118210/fd0aa9aeb65d/PDI3-1-e42-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e9b5/12118210/155f3eb231c5/PDI3-1-e42-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e9b5/12118210/fd0aa9aeb65d/PDI3-1-e42-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e9b5/12118210/155f3eb231c5/PDI3-1-e42-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e9b5/12118210/fd0aa9aeb65d/PDI3-1-e42-g001.jpg

相似文献

[1]
Evaluation of ChatGPT's performance in providing treatment recommendations for pediatric diseases.

Pediatr Discov. 2023-11-20

[2]
Using Artificial Intelligence ChatGPT to Access Medical Information about Chemical Eye Injuries: A Comparative Study.

JMIR Form Res. 2025-6-30

[3]
Evaluating the utility of ChatGPT in addressing conceptual and non-conceptual questions related to urodynamic quality control and trace analysis.

Sci Rep. 2025-6-19

[4]
Can generative artificial intelligence provide accurate medical advice?: a case of ChatGPT versus Congress of Neurological Surgeons management of acute cervical spine and spinal cord injuries clinical guidelines.

Asian Spine J. 2025-3-4

[5]
ChatGPT-supported patient triage with voice commands in the emergency department: A prospective multicenter study.

Am J Emerg Med. 2025-4-17

[6]
Comparison of ChatGPT and Internet Research for Clinical Research and Decision-Making in Occupational Medicine: Randomized Controlled Trial.

JMIR Form Res. 2025-5-20

[7]
Performance of ChatGPT-4o and Four Open-Source Large Language Models in Generating Diagnoses Based on China's Rare Disease Catalog: Comparative Study.

J Med Internet Res. 2025-6-18

[8]
Stench of Errors or the Shine of Potential: The Challenge of (Ir)Responsible Use of ChatGPT in Speech-Language Pathology.

Int J Lang Commun Disord. 2025

[9]
"Dr. AI Will See You Now": How Do ChatGPT-4 Treatment Recommendations Align With Orthopaedic Clinical Practice Guidelines?

Clin Orthop Relat Res. 2024-12-1

[10]
Response to "Letter to the Editor-Exploring the Unknown: Evaluating ChatGPT's Performance in Uncovering Novel Aspects of Plastic Surgery and Identifying Areas for Future Innovation".

Aesthetic Plast Surg. 2024-7-8

引用本文的文献

[1]
Expert evaluation of ChatGPT accuracy and reliability for basic celiac disease frequently asked questions.

Sci Rep. 2025-8-14

[2]
Evaluation of ChatGPT-4 as an Online Outpatient Assistant in Puerperal Mastitis Management: Content Analysis of an Observational Study.

JMIR Med Inform. 2025-7-24

本文引用的文献

[1]
The diagnostic and triage accuracy of the GPT-3 artificial intelligence model: an observational study.

Lancet Digit Health. 2024-8

[2]
ChatGPT in healthcare: A taxonomy and systematic review.

Comput Methods Programs Biomed. 2024-3

[3]
ChatGPT Performs on the Chinese National Medical Licensing Examination.

J Med Syst. 2023-8-15

[4]
Analysis of large-language model versus human performance for genetics questions.

Eur J Hum Genet. 2024-4

[5]
Performance of ChatGPT on the pharmacist licensing examination in Taiwan.

J Chin Med Assoc. 2023-7-1

[6]
Accuracy of Information Provided by ChatGPT Regarding Liver Cancer Surveillance and Diagnosis.

AJR Am J Roentgenol. 2023-10

[7]
Capacity of ChatGPT to Identify Guideline-Based Treatments for Advanced Solid Tumors.

Cureus. 2023-4-21

[8]
The promise and peril of using a large language model to obtain clinical information: ChatGPT performs strongly as a fertility counseling tool with limitations.

Fertil Steril. 2023-9

[9]
Chat Generative Pretrained Transformer Fails the Multiple-Choice American College of Gastroenterology Self-Assessment Test.

Am J Gastroenterol. 2023-12-1

[10]
ChatGPT Answers Common Patient Questions About Colonoscopy.

Gastroenterology. 2023-8

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

推荐工具

医学文档翻译智能文献检索