文献检索文档翻译深度研究
Suppr Zotero 插件Zotero 插件
邀请有礼套餐&价格历史记录

新学期,新优惠

限时优惠:9月1日-9月22日

30天高级会员仅需29元

1天体验卡首发特惠仅需5.99元

了解详情
不再提醒
插件&应用
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
高级版
套餐订阅购买积分包
AI 工具
文献检索文档翻译深度研究
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2025

人工智能在膀胱输尿管反流疾病中的应用:指南依从性的比较研究

Use of Artificial Intelligence in Vesicoureteral Reflux Disease: A Comparative Study of Guideline Compliance.

作者信息

Sarikaya Mehmet, Ozcan Siki Fatma, Ciftci Ilhan

机构信息

Department of Pediatric Surgery, Faculty of Medicine, Selcuk University, Konya 42100, Turkey.

出版信息

J Clin Med. 2025 Mar 30;14(7):2378. doi: 10.3390/jcm14072378.


DOI:10.3390/jcm14072378
PMID:40217829
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11989457/
Abstract

This study aimed to evaluate the compliance of four different artificial intelligence applications (ChatGPT-4.0, Bing AI, Google Bard, and Perplexity) with the American Urological Association (AUA) vesicoureteral reflux (VUR) management guidelines. Fifty-one questions derived from the AUA guidelines were asked of each AI application. Two experienced paediatric surgeons independently scored the responses using a five-point Likert scale. Inter-rater agreement was analysed using the intraclass correlation coefficient (ICC). ChatGPT-4.0, Bing AI, Google Bard, and Perplexity received mean scores of 4.91, 4.85, 4.75 and 4.70 respectively. There was no statistically significant difference between the accuracy of the AI applications ( = 0.223). The inter-rater ICC values were above 0.9 for all platforms, indicating a high level of consistency in scoring. The evaluated AI applications agreed highly with the AUA VUR management guidelines. These results suggest that AI applications may be a potential tool for providing guideline-based recommendations in paediatric urology.

摘要

本研究旨在评估四种不同的人工智能应用程序(ChatGPT-4.0、必应人工智能、谷歌巴德和Perplexity)对美国泌尿外科学会(AUA)膀胱输尿管反流(VUR)管理指南的遵循情况。向每个人工智能应用程序提出了51个源自AUA指南的问题。两名经验丰富的儿科外科医生使用五点李克特量表对回答进行独立评分。使用组内相关系数(ICC)分析评分者间的一致性。ChatGPT-4.0、必应人工智能、谷歌巴德和Perplexity的平均得分分别为4.91、4.85、4.75和4.70。人工智能应用程序的准确性之间没有统计学上的显著差异(=0.223)。所有平台的评分者间ICC值均高于0.9,表明评分具有高度一致性。评估的人工智能应用程序与AUA VUR管理指南高度一致。这些结果表明,人工智能应用程序可能是在小儿泌尿外科提供基于指南的建议的潜在工具。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0957/11989457/eb330771d8b8/jcm-14-02378-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0957/11989457/327ec0d6e503/jcm-14-02378-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0957/11989457/fafcfbfe70f4/jcm-14-02378-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0957/11989457/eb330771d8b8/jcm-14-02378-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0957/11989457/327ec0d6e503/jcm-14-02378-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0957/11989457/fafcfbfe70f4/jcm-14-02378-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0957/11989457/eb330771d8b8/jcm-14-02378-g003.jpg

相似文献

[1]
Use of Artificial Intelligence in Vesicoureteral Reflux Disease: A Comparative Study of Guideline Compliance.

J Clin Med. 2025-3-30

[2]
Comparative analysis of artificial intelligence chatbot recommendations for urolithiasis management: A study of EAU guideline compliance.

Fr J Urol. 2024-7

[3]
Accuracy and Readability of Artificial Intelligence Chatbot Responses to Vasectomy-Related Questions: Public Beware.

Cureus. 2024-8-28

[4]
Battle of the bots: a comparative analysis of ChatGPT and bing AI for kidney stone-related questions.

World J Urol. 2024-10-29

[5]
Harnessing artificial intelligence in bariatric surgery: comparative analysis of ChatGPT-4, Bing, and Bard in generating clinician-level bariatric surgery recommendations.

Surg Obes Relat Dis. 2024-7

[6]
Pilot Testing of a Tool to Standardize the Assessment of the Quality of Health Information Generated by Artificial Intelligence-Based Models.

Cureus. 2023-11-24

[7]
Assessing the Accuracy of Information on Medication Abortion: A Comparative Analysis of ChatGPT and Google Bard AI.

Cureus. 2024-1-2

[8]
The performance of artificial intelligence large language model-linked chatbots in surgical decision-making for gastroesophageal reflux disease.

Surg Endosc. 2024-5

[9]
Performance of Artificial Intelligence Chatbots on Glaucoma Questions Adapted From Patient Brochures.

Cureus. 2024-3-23

[10]
Radiologic Decision-Making for Imaging in Pulmonary Embolism: Accuracy and Reliability of Large Language Models-Bing, Claude, ChatGPT, and Perplexity.

Indian J Radiol Imaging. 2024-7-4

引用本文的文献

[1]
Evaluation of deepseek, gemini, ChatGPT-4o, and perplexity in responding to salivary gland cancer.

BMC Oral Health. 2025-8-23

本文引用的文献

[1]
Artificial intelligence can help individualize Wilms tumor treatment by predicting tumor response to preoperative chemotherapy.

Investig Clin Urol. 2025-1

[2]
Adaptive Treatment of Metastatic Prostate Cancer Using Generative Artificial Intelligence.

Clin Med Insights Oncol. 2025-1-6

[3]
ChatGPT-4o's performance on pediatric Vesicoureteral reflux.

J Pediatr Urol. 2025-4

[4]
Artificial intelligence application in the diagnosis and treatment of bladder cancer: advance, challenges, and opportunities.

Front Oncol. 2024-11-7

[5]
AI-PEDURO - Artificial intelligence in pediatric urology: Protocol for a living scoping review and online repository.

J Pediatr Urol. 2025-4

[6]
Artificial intelligence in urolithiasis: a systematic review of utilization and effectiveness.

World J Urol. 2024-10-17

[7]
Inconsistent advice by ChatGPT influences decision making in various areas.

Sci Rep. 2024-7-10

[8]
Comparative analysis of artificial intelligence chatbot recommendations for urolithiasis management: A study of EAU guideline compliance.

Fr J Urol. 2024-7

[9]
Evaluation of the Impact of ChatGPT on the Selection of Surgical Technique in Bariatric Surgery.

Obes Surg. 2025-1

[10]
The efficacy of artificial intelligence in urology: a detailed analysis of kidney stone-related queries.

World J Urol. 2024-3-14

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

推荐工具

医学文档翻译智能文献检索