• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

与 2020 年儿童舌系带过紧共识声明相比,ChatGPT 的知识比较。

Comparison of ChatGPT knowledge against 2020 consensus statement on ankyloglossia in children.

机构信息

Boston University Chobanian & Avedisian School of Medicine, Boston, MA, USA.

Boston University Chobanian & Avedisian School of Medicine, Boston, MA, USA.

出版信息

Int J Pediatr Otorhinolaryngol. 2024 May;180:111957. doi: 10.1016/j.ijporl.2024.111957. Epub 2024 Apr 16.

DOI:10.1016/j.ijporl.2024.111957
PMID:38640573
Abstract

OBJECTIVE

This paper evaluates ChatGPT's accuracy and consistency in providing information on ankyloglossia, a congenital oral condition. Assessing alignment with expert consensus, the study explores potential implications for patients relying on AI for medical information.

METHODS

Statements from the 2020 clinical consensus statement on ankyloglossia were presented to ChatGPT, and its responses were scored using a 9-point Likert scale. The study analyzed the mean and standard deviation of ChatGPT scores for each statement. Statistical analysis was conducted using Excel.

RESULTS

Among the 63 statements assessed, 67 % of ChatGPT responses closely aligned with expert consensus mean scores. However, 17 % (11/63) were statements in which the ChatGPT mean response was different from the CCS mean by 2.0 or greater, raising concerns about ChatGPT's potential influence in disseminating uncertain or debated medical information. Variations in mean scores highlighted discrepancies, with some statements showing significant deviations from expert opinions.

CONCLUSION

While ChatGPT mirrored medical viewpoints on ankyloglossia, alignment with non-consensus statements raises caution in relying on it for medical advice. Future research should refine AI models, address inaccuracies, and explore diverse user queries for safe integration into medical decision-making. Despite potential benefits, ongoing examination of ChatGPT's power and limitations is crucial, considering its impact on health equity and information access.

摘要

目的

本文评估了 ChatGPT 在提供关于舌系带过紧(一种先天性口腔状况)信息方面的准确性和一致性。评估与专家共识的一致性,探讨了患者依赖 AI 获得医学信息的潜在影响。

方法

将 2020 年舌系带过紧临床共识声明中的陈述提供给 ChatGPT,并使用 9 分李克特量表对其回复进行评分。本研究分析了每个陈述的 ChatGPT 评分的平均值和标准差。使用 Excel 进行统计分析。

结果

在评估的 63 个陈述中,67%的 ChatGPT 回复与专家共识平均得分密切一致。然而,17%(11/63)的陈述中,ChatGPT 的平均回复与 CCS 的平均回复相差 2.0 或更大,这引发了对 ChatGPT 在传播不确定或有争议的医学信息方面的潜在影响的担忧。平均分数的差异突显了差异,一些陈述与专家意见存在显著偏差。

结论

虽然 ChatGPT 反映了舌系带过紧的医学观点,但与非共识陈述的一致性表明,在提供医学建议时应谨慎使用它。未来的研究应该改进 AI 模型,解决不准确的问题,并探索多样化的用户查询,以安全地将其纳入医学决策制定。尽管具有潜在的益处,但需要持续审查 ChatGPT 的能力和局限性,因为它会对健康公平和信息获取产生影响。

相似文献

1
Comparison of ChatGPT knowledge against 2020 consensus statement on ankyloglossia in children.与 2020 年儿童舌系带过紧共识声明相比,ChatGPT 的知识比较。
Int J Pediatr Otorhinolaryngol. 2024 May;180:111957. doi: 10.1016/j.ijporl.2024.111957. Epub 2024 Apr 16.
2
Navigating ChatGPT's alignment with expert consensus on pediatric OSA management.探讨 ChatGPT 在儿科阻塞性睡眠呼吸暂停(OSA)管理方面与专家共识的一致性。
Int J Pediatr Otorhinolaryngol. 2024 Nov;186:112131. doi: 10.1016/j.ijporl.2024.112131. Epub 2024 Oct 15.
3
Assessing the Accuracy of Generative Conversational Artificial Intelligence in Debunking Sleep Health Myths: Mixed Methods Comparative Study With Expert Analysis.评估生成式对话人工智能在破除睡眠健康误区方面的准确性:采用专家分析的混合方法比较研究
JMIR Form Res. 2024 Apr 16;8:e55762. doi: 10.2196/55762.
4
A Comparison of ChatGPT and Expert Consensus Statements on Surgical Site Infection Prevention in High-Risk Paediatric Spine Surgery.ChatGPT与专家共识声明在高危儿科脊柱手术中预防手术部位感染方面的比较
J Pediatr Orthop. 2025 Jan 1;45(1):e72-e75. doi: 10.1097/BPO.0000000000002781. Epub 2024 Aug 30.
5
Performance of ChatGPT on the Chinese Postgraduate Examination for Clinical Medicine: Survey Study.ChatGPT 在临床医学研究生入学考试中的表现:调查研究。
JMIR Med Educ. 2024 Feb 9;10:e48514. doi: 10.2196/48514.
6
An assessment of ChatGPT's responses to frequently asked questions about cervical and breast cancer.评估 ChatGPT 对宫颈癌和乳腺癌常见问题的回答。
BMC Womens Health. 2024 Sep 2;24(1):482. doi: 10.1186/s12905-024-03320-8.
7
Evaluating the Influence of Role-Playing Prompts on ChatGPT's Misinformation Detection Accuracy: Quantitative Study.评估角色扮演提示对 ChatGPT 错误信息检测准确率的影响:定量研究。
JMIR Infodemiology. 2024 Sep 26;4:e60678. doi: 10.2196/60678.
8
ChatGPT's performance in German OB/GYN exams - paving the way for AI-enhanced medical education and clinical practice.ChatGPT在德国妇产科考试中的表现——为人工智能强化医学教育和临床实践铺平道路。
Front Med (Lausanne). 2023 Dec 13;10:1296615. doi: 10.3389/fmed.2023.1296615. eCollection 2023.
9
ChatGPT's Performance in Cardiac Arrest and Bradycardia Simulations Using the American Heart Association's Advanced Cardiovascular Life Support Guidelines: Exploratory Study.ChatGPT 在使用美国心脏协会高级心血管生命支持指南进行心脏骤停和心动过缓模拟中的表现:探索性研究。
J Med Internet Res. 2024 Apr 22;26:e55037. doi: 10.2196/55037.
10
ChatGPT's Response Consistency: A Study on Repeated Queries of Medical Examination Questions.ChatGPT的回答一致性:关于医学考试问题重复查询的研究
Eur J Investig Health Psychol Educ. 2024 Mar 8;14(3):657-668. doi: 10.3390/ejihpe14030043.

引用本文的文献

1
Comparison of responses from different artificial intelligence-powered chatbots regarding the All-on-four dental implant concept.不同人工智能驱动的聊天机器人对全口四颗种植牙概念的回答比较。
BMC Oral Health. 2025 Jun 5;25(1):922. doi: 10.1186/s12903-025-06294-7.