Kapoor Deepika, Garg Deepanshu, Tadakamadla Santosh Kumar
Department of Pedodontics and Preventive Dentistry, Luxmi Bai Institute of Dental Sciences, Baba Farid University of Health Sciences, Punjab, India.
Department of Oral Medicine and Radiology, Luxmi Bai Institute of Dental Sciences, Baba Farid University of Health Sciences, Punjab, India.
Front Oral Health. 2025 Aug 15;6:1652422. doi: 10.3389/froh.2025.1652422. eCollection 2025.
Artificial intelligence (AI) tools such as ChatGPT, Google Gemini, and Microsoft Copilot are increasingly relied upon by parents for immediate guidance on pediatric dental concerns. This study evaluated and compared the response quality of these AI platforms in addressing real-world parental queries related to pediatric dentistry, including early tooth extraction, space maintenance, and the decision to consult a pediatric or a general dentist.
A structured 30-question survey was developed and submitted to each AI model, and their responses were anonymized and assessed by pediatric dental experts using a standardized rubric across five key domains: clinical accuracy, clarity, completeness, relevance, and absence of misleading information.
Statistically significant differences were found across all five domains ( < .001), with ChatGPT consistently achieving the highest scores. Multivariate analysis (MANOVA) confirmed a strong overall effect of the AI model on response quality (Pillai's Trace = 0.892, < .001), supporting ChatGPT's superior performance in providing accurate, relevant, and comprehensive pediatric dental advice.
While AI technologies show potential as clinical decision support systems, their variable performance reinforces the need for expert oversight. Future AI development should focus on optimizing response quality and safety to ensure effective and trustworthy digital health communication for pediatric dental care.
ChatGPT、谷歌Gemini和微软Copilot等人工智能(AI)工具越来越受到家长的依赖,用于获取有关儿童牙科问题的即时指导。本研究评估并比较了这些人工智能平台在处理与儿童牙科相关的现实世界家长问题时的回答质量,这些问题包括早期拔牙、间隙保持以及咨询儿童牙医还是普通牙医的决策。
制定了一份包含30个问题的结构化调查问卷,并提交给每个人工智能模型,其回答进行了匿名处理,由儿童牙科专家使用标准化评分标准在五个关键领域进行评估:临床准确性、清晰度、完整性、相关性以及是否存在误导性信息。
在所有五个领域均发现了具有统计学意义的差异(<0.001),ChatGPT始终获得最高分。多变量分析(MANOVA)证实了人工智能模型对回答质量有很强的总体影响(Pillai迹=0.892,<0.001),支持ChatGPT在提供准确、相关和全面的儿童牙科建议方面的卓越表现。
虽然人工智能技术作为临床决策支持系统显示出潜力,但其表现参差不齐,这强化了专家监督的必要性。未来的人工智能开发应专注于优化回答质量和安全性,以确保为儿童牙科护理提供有效且值得信赖的数字健康沟通。